Multi-Domain Long-Tailed Recognition (MDLT)
This repository contains the implementation for paper: On Multi-Domain Long-Tailed Recognition, Imbalanced Domain Generalization and Beyond (ECCV 2022).
It is also a (living) PyTorch suite containing benchmark datasets and algorithms for Multi-Domain Long-Tailed Recognition (MDLT). Currently we support 8 MDLT datasets (3 synthetic + 5 real), as well as ~20 algorithms that span different learning strategies. Feel free to send us a PR to add your algorithm / dataset for MDLT!
Multi-Domain Long-Tailed Recognition (MDLT) aims to learn from multi-domain imbalanced data, address label imbalance, domain shift, and divergent label distributions across domains, and generalize to all domain-class pairs.
MDLT: From Single- to Multi-Domain Imbalanced Learning
Existing studies on data imbalance focus on single-domain settings, i.e., samples are from the same data distribution. However, natural data can originate from distinct domains, where a minority class in one domain could have abundant instances from other domains. We systematically investigate Multi-Domain Long-Tailed Recognition (MDLT), which learns from multi-domain imbalanced data, addresses label imbalance, domain shift, and divergent label distributions across domains, and generalizes to all domain-class pairs.
We develop the domain-class transferability graph, and show that such transferability governs the success of learning in MDLT. We then propose BoDA, a theoretically grounded learning strategy that tracks the upper bound of transferability statistics, and ensures balanced alignment and calibration across imbalanced domain-class distributions. We curate MDLT benchmark datasets based on widely-used multi-domain datasets, and benchmark ~20 algorithms that span different learning strategies for MDLT.
Beyond MDLT: Domain Generalization under Data Imbalance
Further, as a byproduct, we demonstrate that BoDA strengthens Domain Generalization (DG) algorithms, and consistently improves the results on DG benchmarks. Note that all current standard DG benchmarks naturally exhibit heavy class imbalance within domains and label distributions shift across domains, confirming that data imbalance is an intrinsic problem in DG, but has yet been overlooked by past works.
The results shed light on how label imbalance can affect out-of-distribution generalization, and highlight the importance of integrating label imbalance into practical DG algorithm design.
Getting Started
Installation
Prerequisites
- Download the original datasets, and place them in your
data_path
python -m mdlt.scripts.download --data_dir <data_path>
- Place the
.csv
files of train/val/test splits for each MDLT dataset (provided inmdlt/dataset/split/
) in the corresponding dataset folder under yourdata_path
Dependencies
- PyTorch (>=1.4, tested on 1.4 / 1.9)
- pandas
- TensorboardX
Code Overview
Main Files
train.py
: main training scriptsweep.py
: launch a sweep with all selected algorithms (provided inmdlt/learning/algorithms.py
) on all real MDLT datasets (VLCS-MLT, PACS-MLT, OfficeHome-MLT, TerraInc-MLT, DomainNet-MLT)sweep_synthetic.py
: launch a sweep with all selected algorithms on the synthetic MDLT dataset (Digits-MLT)collect_results.py
: collect sweep results to automatically generate result tables (as in the paper)eval_best_hparam.py
&eval_checkpoint.py
: scripts for evaluating trained models
Main Arguments
- train.py:
--dataset
: name of chosen MDLT dataset--algorithm
: choose algorithm used for running--data_dir
: data path--output_dir
: output path--output_folder_name
: output folder name (underoutput_dir
) for the current run--hparams_seed
: seed for different hyper-parameters--seed
: seed for different runs--selected_envs
: train on selected envs (only used for Digits-MLT)--imb_type
&--imb_factor
: arguments for customized Digits-MLT label distributions--stage1_folder
&--stage1_algo
: arguments for two-stage algorithms
- sweep.py:
--n_hparams
: how many hparams to run for each <dataset, algorithm> pair--best_hp
&--n_trials
: after sweeping hparams, fix best hparam and run trials with different seeds
Usage
Train a single model
python -m mdlt.train --algorithm <algo> --dataset <dset> --output_folder_name <output_folder_name> --data_dir <data_path> --output_dir <output_path>
Train a model using 2-stage (second stage classifier learning)
python -m mdlt.train --algorithm CRT --dataset <dset> --output_folder_name <output_folder_name> --data_dir <data_path> --output_dir <output_path> --stage1_folder <stage1_model_folder> --stage1_algo <stage1_algo>
Note that for stage1_algo
& stage1_folder
Forward-LT
and imbalance ratio 0.01
Train a model on Digits-MLT, with imbalance type all python -m mdlt.train --algorithm <algo> --dataset ImbalancedDigits \
--imb_type eee \
--imb_factor 0.01 \
--selected_envs 1 2
Note that for Digits-MLT, we additionally provide MNIST
as another domain.
To maintain the same setting as in paper (2 domains), you only need to set selected_envs
to be 1 2
as above
Launch a sweep with different hparams
python -m mdlt.sweep launch --algorithms <...> --dataset <...> --n_hparams <num_of_hparams> --n_trials 1
Launch a sweep after fixing hparam with different seeds
python -m mdlt.sweep launch --algorithms <...> --dataset <...> --best_hp --input_folder <...> --n_trials <num_of_trials>
Collect the results of your sweep
python -m mdlt.scripts.collect_results --input_dir <...>
Evaluate the best hparam model for a <dataset, algo> pair
python -u -m mdlt.evaluate.eval_best_hparam --algorithm <...> --dataset <...> --data_dir <...> --output_dir <...> --folder_name <...>
Evaluate a trained checkpoint
python -u -m mdlt.evaluate.eval_checkpoint --algorithm <...> --dataset <...> --data_dir <...> --checkpoint <...>
Reproduced Benchmarks and Model Zoo
Model | VLCS-MLT | PACS-MLT | OfficeHome-MLT | TerraInc-MLT | DomainNet-MLT |
---|---|---|---|---|---|
BoDA (r) | 76.9 / model | 97.0 / model | 81.5 / model | 78.6 / model | 60.1 / model |
BoDA (r,c) | 77.3 / model | 97.2 / model | 82.3 / model | 82.3 / model | 61.7 / model |
Updates
- [10/2022] Check out the Oral talk video (10 mins) for our ECCV paper.
- [07/2022] We create a Blog post for this work (version in Chinese is also available here). Check it out for more details!
- [07/2022] Paper accepted to ECCV 2022. We have released the code and models.
- [03/2022] arXiv version posted. The code is currently under cleaning. Please stay tuned for updates.
Acknowledgements
This code is partly based on the open-source implementations from DomainBed.
Citation
If you find this code or idea useful, please cite our work:
@inproceedings{yang2022multi,
title={On Multi-Domain Long-Tailed Recognition, Imbalanced Domain Generalization and Beyond},
author={Yang, Yuzhe and Wang, Hao and Katabi, Dina},
booktitle={European Conference on Computer Vision (ECCV)},
year={2022}
}
Contact
If you have any questions, feel free to contact us through email ([email protected]) or Github issues. Enjoy!