Publication
Scaling Probabilistic Circuits via Data Partitioning
Jonas Seng; Florian Peter Busch; Pooja Prasad; Devendra Singh Dhami; Martin Mundt; Kristian Kersting
In: Computing Research Repository eprint Journal (CoRR), Vol. abs/2503.08141, Pages 1-15, Computing Research Repository, 2025.
Abstract
Probabilistic circuits (PCs) enable us to learn joint
distributions over a set of random variables and to
perform various probabilistic queries in a tractable
fashion. Though the tractability property allows
PCs to scale beyond non-tractable models such as
Bayesian Networks, scaling training and inference
of PCs to larger, real-world datasets remains chal-
lenging. To remedy the situation, we show how
PCs can be learned across multiple machines by re-
cursively partitioning a distributed dataset, thereby
unveiling a deep connection between PCs and fed-
erated learning (FL). This leads to federated cir-
cuits (FCs)—a novel and flexible federated learn-
ing (FL) framework that (1) allows one to scale
PCs on distributed learning environments (2) train
PCs faster and (3) unifies for the first time horizon-
tal, vertical, and hybrid FL in one framework by
re-framing FL as a density estimation problem over
distributed datasets. We demonstrate FC’s capabil-
ity to scale PCs on various large-scale datasets.
Also, we show FC’s versatility in handling hor-
izontal, vertical, and hybrid FL within a unified
framework on multiple classification tasks.
