This research proposes a novel sampling algorithm for speech super-resolution that leverages the power of variational diffusion models (VDMs) to reconstruct high-resolution speech from low-resolution audio, achieving state-of-the-art results and demonstrating robustness against different downsampling methods.
Wave-U-Mamba is an efficient and effective end-to-end framework for speech super-resolution that directly generates high-resolution speech waveforms from low-resolution inputs, outperforming existing state-of-the-art models.