Generative Adversarial Networks and Perceptual Losses for Video Super-Resolution. Lucas, A., Lopez-Tapia, S., Molina, R., & Katsaggelos, A. K. IEEE Transactions on Image Processing, 28(7):3312–3327, jul, 2019.
Generative Adversarial Networks and Perceptual Losses for Video Super-Resolution [link]Paper  doi  abstract   bibtex   
Video super-resolution (VSR) has become one of the most critical problems in video processing. In the deep learning literature, recent works have shown the benefits of using adversarial-based and perceptual losses to improve the performance on various image restoration tasks; however, these have yet to be applied for video super-resolution. In this paper, we propose a generative adversarial network (GAN)-based formulation for VSR. We introduce a new generator network optimized for the VSR problem, named VSRResNet, along with new discriminator architecture to properly guide VSRResNet during the GAN training. We further enhance our VSR GAN formulation with two regularizers, a distance loss in feature-space and pixel-space, to obtain our final VSRResFeatGAN model. We show that pre-training our generator with the mean-squared-error loss only quantitatively surpasses the current state-of-the-art VSR models. Finally, we employ the PercepDist metric to compare the state-of-the-art VSR models. We show that this metric more accurately evaluates the perceptual quality of SR solutions obtained from neural networks, compared with the commonly used PSNR/SSIM metrics. Finally, we show that our proposed model, the VSRResFeatGAN model, outperforms the current state-of-the-art SR models, both quantitatively and qualitatively.
@article{Alice2019c,
abstract = {Video super-resolution (VSR) has become one of the most critical problems in video processing. In the deep learning literature, recent works have shown the benefits of using adversarial-based and perceptual losses to improve the performance on various image restoration tasks; however, these have yet to be applied for video super-resolution. In this paper, we propose a generative adversarial network (GAN)-based formulation for VSR. We introduce a new generator network optimized for the VSR problem, named VSRResNet, along with new discriminator architecture to properly guide VSRResNet during the GAN training. We further enhance our VSR GAN formulation with two regularizers, a distance loss in feature-space and pixel-space, to obtain our final VSRResFeatGAN model. We show that pre-training our generator with the mean-squared-error loss only quantitatively surpasses the current state-of-the-art VSR models. Finally, we employ the PercepDist metric to compare the state-of-the-art VSR models. We show that this metric more accurately evaluates the perceptual quality of SR solutions obtained from neural networks, compared with the commonly used PSNR/SSIM metrics. Finally, we show that our proposed model, the VSRResFeatGAN model, outperforms the current state-of-the-art SR models, both quantitatively and qualitatively.},
archivePrefix = {arXiv},
arxivId = {1806.05764},
author = {Lucas, Alice and Lopez-Tapia, Santiago and Molina, Rafael and Katsaggelos, Aggelos K.},
doi = {10.1109/TIP.2019.2895768},
eprint = {1806.05764},
issn = {1057-7149},
journal = {IEEE Transactions on Image Processing},
keywords = {Artificial neural networks,image generation,image resolution,video signal processing},
month = {jul},
number = {7},
pages = {3312--3327},
pmid = {30714918},
title = {{Generative Adversarial Networks and Perceptual Losses for Video Super-Resolution}},
url = {https://ieeexplore.ieee.org/document/8629024/},
volume = {28},
year = {2019}
}

Downloads: 0