• Stars
    star
    641
  • Rank 70,212 (Top 2 %)
  • Language
    Python
  • License
    Apache License 2.0
  • Created about 3 years ago
  • Updated over 1 year ago

Reviews

There are no reviews yet. Be the first to send feedback to the community and the maintainers!

Repository Details

Official repository of "Investigating Tradeoffs in Real-World Video Super-Resolution"

RealBasicVSR (CVPR 2022)

[Paper]

This is the official repository of "Investigating Tradeoffs in Real-World Video Super-Resolution, arXiv". This repository contains codes, colab, video demos of our work.

Authors: Kelvin C.K. Chan, Shangchen Zhou, Xiangyu Xu, Chen Change Loy, Nanyang Technological University

Acknowedgement: Our work is built upon MMEditing. The code will also appear in MMEditing soon. Please follow and star this repository and MMEditing!

Feel free to ask questions. I am currently working on some other stuff but will try my best to reply. If you are also interested in BasicVSR++, which is also accepted to CVPR 2022, please don't hesitate to star!

News

  • 11 Mar 2022: Integrated into Huggingface Spaces 🤗 using Gradio. Try out the Web Demo: Hugging Face Spaces
  • 3 Mar 2022: Our paper has been accepted to CVPR 2022
  • 4 Jan 2022: Training code released
  • 2 Dec 2021: Colab demo released google colab logo
  • 29 Nov 2021: Test code released
  • 25 Nov 2021: Initialize with video demos

Table of Content

  1. Video Demos
  2. Code
  3. VideoLQ Dataset
  4. Citations

Video Demos

The videos have been compressed. Therefore, the results are inferior to that of the actual outputs.

output.mp4
output.mp4
output.mp4
output.mp4

Code

Installation

  1. Install PyTorch and torchvision following the official instructions, e.g.,
conda install pytorch==1.7.1 torchvision==0.8.2 torchaudio==0.7.2 cudatoolkit=10.1 -c pytorch
  1. Install mim and mmcv-full
pip install openmim
mim install mmcv-full
  1. Install mmedit
pip install mmedit

Inference

  1. Download the pre-trained weights to checkpoints/. (Dropbox / Google Drive / OneDrive)

  2. Run the following command:

python inference_realbasicvsr.py ${CONFIG_FILE} ${CHECKPOINT_FILE} ${INPUT_DIR} ${OUTPUT_DIR} --max-seq-len=${MAX_SEQ_LEN} --is_save_as_png=${IS_SAVE_AS_PNG}  --fps=${FPS}

This script supports both images and videos as inputs and outputs. You can simply change ${INPUT_DIR} and ${OUTPUT_DIR} to the paths corresponding to the video files, if you want to use videos as inputs and outputs. But note that saving to videos may induce additional compression, which reduces output quality.

For example:

  1. Images as inputs and outputs
python inference_realbasicvsr.py configs/realbasicvsr_x4.py checkpoints/RealBasicVSR_x4.pth data/demo_000 results/demo_000
  1. Video as input and output
python inference_realbasicvsr.py configs/realbasicvsr_x4.py checkpoints/RealBasicVSR_x4.pth data/demo_001.mp4 results/demo_001.mp4 --fps=12.5

Crop REDS dataset into sub-images

We crop the REDS dataset into sub-images for faster I/O. Please follow the instructions below:

  1. Put the original REDS dataset in ./data
  2. Run the following command:
python crop_sub_images.py --data-root ./data/REDS  --scales 4

Training

The training is divided into two stages:

  1. Train a model without perceptual loss and adversarial loss using realbasicvsr_wogan_c64b20_2x30x8_lr1e-4_300k_reds.py.
mim train mmedit configs/realbasicvsr_wogan_c64b20_2x30x8_lr1e-4_300k_reds.py --gpus 8 --launcher pytorch
  1. Finetune the model with perceptual loss and adversarial loss using realbasicvsr_c64b20_1x30x8_lr5e-5_150k_reds.py. (You may want to replace load_from in the configuration file with your checkpoints pre-trained at the first stage
mim train mmedit configs/realbasicvsr_c64b20_1x30x8_lr5e-5_150k_reds.py --gpus 8 --launcher pytorch

Note: We use UDM10 with bicubic downsampling for validation. You can download it from here.

Generating Video Demo

Assuming you have created two sets of images (e.g. input vs output), you can use generate_video_demo.py to generate a video demo. Note that the two sets of images must be of the same resolution. An example has been provided in the code.

VideoLQ Dataset

You can download the dataset using Dropbox / Google Drive / OneDrive.

Citations

@inproceedings{chan2022investigating,
  author = {Chan, Kelvin C.K. and Zhou, Shangchen and Xu, Xiangyu and Loy, Chen Change},
  title = {Investigating Tradeoffs in Real-World Video Super-Resolution},
  booktitle = {IEEE Conference on Computer Vision and Pattern Recognition},
  year = {2022}
}