The paper studies the generalization capabilities of message passing neural networks (MPNNs) in a more realistic setting compared to previous work. The key modifications are:
The authors propose a generative model for graph-signals based on a mixture of graphons, where each class is associated with a unique graphon. They derive non-asymptotic generalization bounds for supervised graph classification tasks using MPNNs in this more realistic setting. The bounds show that as the average number of nodes in the graphs increases, the generalization error decreases. This implies that MPNNs with higher complexity than the size of the training set can still generalize effectively, as long as the graphs are sufficiently large.
The theoretical results are supported by numerical experiments, which demonstrate that the proposed bounds are significantly tighter than existing bounds.
На другой язык
из исходного контента
arxiv.org
Дополнительные вопросы