Skip to content

Fanghua-Yu/OSRT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

(CVPR2023) OSRT: Omnidirectional Image Super-Resolution with Distortion-aware Transformer [PDF]

Fanghua Yu, Xintao Wang, Mingdeng Cao, Gen Li, Ying Shan, Chao Dong

XPixel Group
Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences
Applied Research Center (ARC), Tencent PCG
The University of Tokyo
Platform Technologies, Tencent Online Video
Shanghai AI Lab


Ground-Truth (X4) EDSR SwinIR SwinIR-AUG OSRT-AUG (ours)

Introduction

Fisheye Downsampling

In practice, ODIs are acquired by the fisheye lens and stored in ERP. Given that the low-resolution issue in real-world scenarios is caused by insufficient sensor precision and density, the downsampling process should be applied to original-formatted images before converting into other storage types. Thus, to be conformed with real-world imaging processes, we propose to apply uniform bicubic downsampling on Fisheye images, which are the original format of ODIs. The new downsampling process (called Fisheye downsampling) applies uniform bicubic downsam-pling on Fisheye images before converting them to ERP im-ages. Our Fisheye downsampling is more conducive to ex-ploring the geometric property of ODIs.

We provide implementation codes of Fisheye Downsampling as an independent downsampling kernel for ODIs:

from PIL import Image
from torchvision import transforms
from odisr.utils.erp_downsample import erp_downsample_fisheye
given_path = ''
save_path = ''
scale = 4
erp_hr = transforms.ToTensor()(Image.open(given_path).convert('RGB'))
erp_lr = erp_downsample_fisheye(erp_hr, scale)
transforms.ToPILImage()(erp_lr).save(save_path)

OSRT

We propose the first Transformer for Omnidirectional image Super-Resolution (OSRT), and incorporate geometric prop-erties in a distortion-aware manner. Specifically, to mod-ulate distorted feature maps, we implement feature-level warping, in which offsets are learned from latitude condi-tions. In OSRT, we introduce two dedicated blocks to adapt latitude-related distortion: distortion-aware attention block (DAAB), and distortion-aware convolution block (DACB). DAAB and DACB are designed to perform distortion mod-ulation in arbitrary Transformers and ConvNets. These two blocks can directly replace the multi-head self-attention block and convolution layer, respectively. The benefit of DAAB and DACB can be further improved when being in-serted into the same backbone network. OSRT outperforms previous methods by about 0.2dB on PSNR.

Pseudo ODI Augmentation

The largest ODIs dataset contains only 1K images, which cannot provide enough diversity for training Transformers. Given that acquiring ODIs requires expensive equipment and tedious work, we propose to gen-erate distorted ERP samples from plain images for data aug-mentation. In practice, we regard a plain image as a sampled perspective, and project it back to the ERP format.


Environment

Installation

pip install -r requirements.txt

Data Preparation

  1. Download lau dataset here.
mkdir datasets
ln -s YOUR_LAU_DATASET_PATH datasets/lau_dataset
  1. Clean dataset.
python odisr/utils/make_clean_lau_dataset.py
  1. Crop training patches
python odisr/utils/extract_subimage.py
  1. Build DF2K-ODI augmentation dataset from DF2K (DIV2K/ Flickr2K)
ln -s YOUR_DIV2K_TRAINSET_PATH datasets/DIV2K_train_HR
ln -s YOUR_FLICKR2K_TRAINSET_PATH datasets/Flickr2K_HR
python odisr/utils/make_augmentation_dataset.py

As DF2K-ODI is already consisted of image patches, cropping patches for lau datasets is NECESSARY when training with this dataset. Maximum gt_size for DF2K-ODI is 256.

Training

We provide training options of OSRT-full (10M) and OSRT-light(1M) on X2, X4, X8 and X16 SR.

All comparison models on X4 SR is available.

Training OSRT-full and SwinIR:

CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python -m torch.distributed.launch --nproc_per_node=8 --master_port=7777 train.py -opt ./options/train/*.yml --launcher pytorch

Training OSRT-light and other baseline models:

CUDA_VISIBLE_DEVICES=0,1,2,3 python -m torch.distributed.launch --nproc_per_node=4 --master_port=7777 train.py -opt ./options/train/*.yml --launcher pytorch

Testing

Pretrained models can be downloaded here.

Testing OSRT-full and OSRT-light:

ln -s YOUR_PRETRAIN_MODEL_PATH pretrained_models
CUDA_VISIBLE_DEVICES=0 python test.py -opt ./options/test/*.yml

Citation

@article{yu2023osrt,
  title={OSRT: Omnidirectional Image Super-Resolution with Distortion-aware Transformer},
  author={Yu, Fanghua and Wang, Xintao and Cao, Mingdeng and Li, Gen and Shan, Ying and Dong, Chao},
  journal={arXiv preprint arXiv:2302.03453},
  year={2023}
}

License and Acknowledgement

This project is released under the MIT license. The codes are heavily based on BasicSR. Please also follow their licenses. Thanks for their awesome works.

About

Official code of OSRT: Omnidirectional Image Super-Resolution with Distortion-aware Transformer

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages