The paper introduces Spectral Convolutional Conditional Neural Processes (SConvCNPs), a new addition to the Conditional Neural Process (CNP) family. SConvCNPs aim to address the limitations of ConvCNPs, which rely on local discrete kernels in their convolution layers, by incorporating Fourier Neural Operators (FNOs) to perform global convolution.
The key highlights are:
Conditional Neural Processes (CNPs) use neural networks to parameterize stochastic processes, providing well-calibrated predictions and simple maximum-likelihood training.
ConvCNPs, a variant of CNPs, utilize convolution to introduce translation equivariance as an inductive bias. However, their reliance on local discrete kernels can pose challenges in capturing long-range dependencies and complex patterns, especially with limited and irregularly sampled observations.
SConvCNPs leverage the formulation of FNOs to perform global convolution, allowing for more efficient representation of functions in the frequency domain.
Experiments on synthetic one-dimensional regression tasks demonstrate that SConvCNPs either match or outperform the performance of baseline models, including Vanilla CNP, Attentive CNP, and ConvCNP, particularly in scenarios with periodic underlying functions.
The use of global convolution in SConvCNPs provides a more robust representation of underlying patterns by considering information collectively, as evidenced by the superior fits produced by SConvCNPs compared to the baselines.
다른 언어로
소스 콘텐츠 기반
arxiv.org
더 깊은 질문