Networked Information Aggregation for Binary Classification
Network depth is the bottleneck for logistic regression in distributed settings, new proof shows.
A team of researchers from academia has published a theoretical analysis of networked binary classification on directed acyclic graphs (DAGs). In their model, each agent sees only a subset of feature columns from a shared dataset. Agents act sequentially: each receives prediction columns from its parents, augments its local features, fits a logistic predictor by minimizing binary cross-entropy (BCE), and forwards its prediction to downstream neighbors. The central question is whether this sequential distributed training achieves information aggregation—i.e., whether some agent attains small excess loss compared to the best logistic predictor trained on all features. Prior work by Kearns, Roth, and Ryu (SODA 2026) answered this for linear regression under squared loss, but their proof relied on quadratic structure that does not transfer to BCE with a logistic link.
The new analysis introduces a sequential logit-passing protocol and proves two main results. First, on depth-D paths, if every M consecutive agents collectively observe all features, the excess loss is bounded by O(M/√D). Second, they prove a lower bound of Ω(k/D), where k is the feature dimension, showing instances where performance degrades linearly with depth. Together, these bounds identify network depth as a fundamental bottleneck for information aggregation in logistic regression. The paper was accepted to the 43rd International Conference on Machine Learning (ICML 2026), signaling its theoretical significance for distributed machine learning systems.
- Excess loss upper bound of O(M/√D) on depth-D paths under the condition that every M consecutive agents collectively see all features.
- Matching lower bound of Ω(k/D) where k is the feature dimension, proving depth is a fundamental bottleneck.
- Paper extends prior networked learning results from linear regression to classification using binary cross-entropy, accepted at ICML 2026.
Why It Matters
Distributed ML systems must account for network depth to avoid degraded classification accuracy in sequential training pipelines.