The paper presents WaveMixSR-V2, an enhanced version of the previous WaveMixSR model for single image super-resolution (SISR). The key improvements include:
Multi-stage Design: Instead of a single upsampling layer, WaveMixSR-V2 uses a series of 2x SR blocks to progressively increase the resolution, allowing better refinement of details at each step.
PixelShuffle: The transposed convolution in the original WaveMixSR blocks is replaced with a PixelShuffle operation followed by a convolution. This significantly reduces the number of parameters and computational cost while avoiding checkerboard artifacts.
The experiments demonstrate that WaveMixSR-V2 outperforms other state-of-the-art methods, including WaveMixSR, SwinIR, and HAT, on the BSD100 dataset for both 2x and 4x super-resolution. Notably, WaveMixSR-V2 achieves these results using less than half the number of parameters and computations compared to the previous best model, WaveMixSR.
Additionally, the paper reports that WaveMixSR-V2 exhibits higher parameter efficiency, lower latency, and higher throughput compared to WaveMixSR, making it one of the most efficient models for super-resolution tasks.
The authors also conducted ablation studies to analyze the impact of different loss functions, input resolutions, and the inclusion of Gaussian noise. The results suggest that the WaveMixSR-V2 architecture's focus on low-frequency components may not align well with the objectives of GAN training, leading to limited benefits from the GAN approach.
翻译成其他语言
从原文生成
arxiv.org
更深入的查询