Wavelet Attention Embedding Networks for Video Super-Resolution
Citations

WEB OF SCIENCE

1
Citations

SCOPUS

16

초록

Recently, Video super-resolution (VSR) has become more crucial as the resolution of display has been grown. The majority of deep learning-based VSR methods combine the convolutional neural networks (CNN) with motion compensation or alignment module to estimate a high-resolution (HR) frame from low-resolution (LR) frames. However, most of the previous methods deal with the spatial features equally and may result in the misaligned temporal features by the pixel-based motion compensation and alignment module. It can lead to the damaging effect on the accuracy of the estimated HR feature. In this paper, we propose a wavelet attention embedding network (WAEN), including wavelet embedding network (WENet) and attention embedding network (AENet), to fully exploit the spatio-temporal informative features. The WENet is operated as a spatial feature extractor of individual low and high-frequency information based on 2-D Haar discrete wavelet transform. The meaningful temporal feature is extracted in the AENet through utilizing the weighted attention map between frames. Experimental results verify that the proposed method achieves superior performance compared with state-of-the-art methods.

제목
Wavelet Attention Embedding Networks for Video Super-Resolution
저자
Choi, Young-JuLee, Young-WoonKim, Byung-Gyu
DOI
10.1109/ICPR48806.2021.9412623
발행일
2021-01
유형
Proceedings Paper
저널명
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR)
페이지
7314 ~ 7320