The paper proposes a unified framework for integrating diverse types of domain knowledge into the parameter learning of Probabilistic Circuits (PCs). PCs are an efficient framework for representing and learning complex probability distributions, but they often struggle with limited and noisy data, similar to deep generative models.
The key contributions are:
Developing a unified mathematical framework that allows encoding different types of domain knowledge as probabilistic constraints, including generalization, monotonicity, context-specific independence, class imbalance, synergy, and privileged information.
Formulating the knowledge-intensive parameter learning of PCs as a constrained optimization problem, where the domain constraints are seamlessly incorporated into the maximum likelihood objective.
Empirically validating the effectiveness of the proposed approach on several benchmark and real-world datasets, demonstrating that incorporating domain knowledge can significantly improve the generalization performance and robustness of PCs, especially in data-scarce and noisy settings.
The experiments show that the framework can faithfully integrate diverse forms of domain knowledge, leading to superior performance compared to purely data-driven approaches. The approach is also shown to be robust to noisy or redundant advice from domain experts.
A otro idioma
del contenido fuente
arxiv.org
Ideas clave extraídas de
by Athresh Kara... a las arxiv.org 05-07-2024
https://arxiv.org/pdf/2405.02413.pdfConsultas más profundas