Arbitrary-Scale Video Super-resolution Guided by Dynamic Context
DOI:
https://doi.org/10.1609/aaai.v38i3.28003Keywords:
CV: Low Level & Physics-based Vision, CV: ApplicationsAbstract
We propose a Dynamic Context-Guided Upsampling (DCGU) module for video super-resolution (VSR) that leverages temporal context guidance to achieve efficient and effective arbitrary-scale VSR. While most VSR research focuses on backbone design, the importance of the upsampling part is often overlooked. Existing methods rely on pixelshuffle-based upsampling, which has limited capabilities in handling arbitrary upsampling scales. Recent attempts to replace pixelshuffle-based modules with implicit neural function-based and filter-based approaches suffer from slow inference speeds and limited representation capacity, respectively. To overcome these limitations, our DCGU module predicts non-local sampling locations and content-dependent filter weights, enabling efficient and effective arbitrary-scale VSR. Our proposed multi-granularity location search module efficiently identifies non-local sampling locations across the entire low-resolution grid, and the temporal bilateral filter modulation module integrates content information with the filter weight to enhance textual details. Extensive experiments demonstrate the superiority of our method in terms of performance and speed on arbitrary-scale VSR.Downloads
Published
2024-03-24
How to Cite
Huang, C., Li, J., Chu, L., Liu, D., & Lu, Y. (2024). Arbitrary-Scale Video Super-resolution Guided by Dynamic Context. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2294-2302. https://doi.org/10.1609/aaai.v38i3.28003
Issue
Section
AAAI Technical Track on Computer Vision II