The paper studies the generalization capabilities of message passing neural networks (MPNNs) in a more realistic setting compared to previous work. The key modifications are:
The authors propose a generative model for graph-signals based on a mixture of graphons, where each class is associated with a unique graphon. They derive non-asymptotic generalization bounds for supervised graph classification tasks using MPNNs in this more realistic setting. The bounds show that as the average number of nodes in the graphs increases, the generalization error decreases. This implies that MPNNs with higher complexity than the size of the training set can still generalize effectively, as long as the graphs are sufficiently large.
The theoretical results are supported by numerical experiments, which demonstrate that the proposed bounds are significantly tighter than existing bounds.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Sohir Maskey... lúc arxiv.org 04-05-2024
https://arxiv.org/pdf/2404.03473.pdfYêu cầu sâu hơn