Disentangled Non-Local Neural Networks
By Minghao Yin, Zhuliang Yao, Yue Cao, Xiu Li, Zheng Zhang, Stephen Lin, Han Hu.
We provide related codes and configuration files to reproduce the "Disentangled Non-Local Neural Networks" on Cityscapes and ADE20K semantic segmentation. Our code is adapted from torchcv.
Introduction
The non-local block models long-range dependency in many computer vision tasks. Our paper first studies the non-local block in depth, where we find that its attention computation can be split into pairwise term and unary term. However, the two terms are tightly coupled in the non-local block, which hinders the learning of each. Based on that, we developed the disentangled non-local block, where the two terms are decoupled to facilitate learning for both terms. Disentangled non-local block outperforms non-local block in various tasks.
The disentangled Non-local block composes of a whitened pairwise term accounting for the relationship between two pixels and a unary term representing the saliency of every pixel. The two terms are clearly decoupled by using independent Softmax functions and embedding matrices. As shown in the attention maps, the whitened pairwise term learns clear within-region clues while the unary term learns salient boundaries.
Citation
@article{yin2020disentangled,
title={Disentangled Non-Local Neural Networks},
author={Yin, Minghao and Yao, Zhuliang and Cao, Yue and Li, Xiu and Zhang, Zheng and Lin, Stephen and Hu, Han},
journal={arXiv preprint arXiv:2006.06668},
year={2020}
}
Install
Requirements
- Linux or macOS (Windows is not currently officially supported)
- Python 3.6+
- PyTorch 1.0+
- CUDA 9.0+
a. Create a conda virtual environment and activate it.
conda create -n dnlnet python=3.6 -y
conda activate dnlnet
b. Install PyTorch and torchvision following the official instructions, e.g.,
conda install pytorch torchvision -c pytorch
c. Clone the repository.
git clone https://github.com/yinmh17/DNL-Networks.git
cd DNL-Networks
d.Install build requirements
pip3 install -r requirements.txt
cd exts
sh make.sh
Train
Before training, you need to preprocess your datasets following the instruction in datasets/seg/preprocess. Download ImageNet pretrained model 3x3-Resnet101
You can use following commands to train your model.
bash ./scripts/seg/{DATASET}/{SCRIPT_FILE_NAME}.sh train {TAG} {DATA_DIR} {PRETRAINED_MODEL}
Optional arguments:
DATASET
: Name of the dataset, for example: "cityscapes", "ade20k".SCRIPT_FILE_NAME
: Name of the bash file you are going to run. Training settings are contained in the bash file.DATA_DIRECTION
: Directory of the preprocessed dataset.PRETRAINED_MODEL
: Directory of your pretrained model.
Take DNL network training on Cityscapes dataset for example, you need to run the following commands.
bash ./scripts/seg/cityscapes/run_fs_res101_nonlocalnowd_ln_cityscapes_seg.sh train tag /torchcv/data/cityscapes ./pretrained_models/3x3resnet101-imagenet.pth
Inference
After training, you can use following commands to validate your model.
bash ./scripts/seg/{DATASET}/{SCRIPT_FILE_NAME}.sh val {TAG} {DATA_DIR}
Take DNL network validation on Cityscapes dataset for example:
bash ./scripts/seg/cityscapes/run_fs_res101_nonlocalnowd_ln_cityscapes_seg.sh val tag /torchcv/data/cityscapes