Inference with Minimal Communication: A Decision-Theoretic Variational Approach

Abstract

Given a directed graphical model with binary-valued hidden nodes and real-valued noisy observations, consider deciding upon the maximum a-posteriori (MAP) or the maximum posterior-marginal (MPM) assignment under the restriction that each node broadcasts only to its children exactly one single-bit message. We present a variational formulation, viewing the processing rules local to all nodes as degrees-of-freedom, that minimizes the loss in expected (MAP or MPM) performance subject to such online communication constraints. The approach leads to a novel message-passing algorithm to be executed offline, or before observations are realized, which mitigates the performance loss by iteratively coupling all rules in a manner implicitly driven by global statistics. We also provide (i) illustrative examples, (ii) assumptions that guarantee convergence and efficiency and (iii) connections to active research areas.

Cite

Text

Kreidl and Willsky. "Inference with Minimal Communication: A Decision-Theoretic Variational Approach." Neural Information Processing Systems, 2005.

Markdown

[Kreidl and Willsky. "Inference with Minimal Communication: A Decision-Theoretic Variational Approach." Neural Information Processing Systems, 2005.](https://mlanthology.org/neurips/2005/kreidl2005neurips-inference/)

BibTeX

@inproceedings{kreidl2005neurips-inference,
  title     = {{Inference with Minimal Communication: A Decision-Theoretic Variational Approach}},
  author    = {Kreidl, O. P. and Willsky, Alan S.},
  booktitle = {Neural Information Processing Systems},
  year      = {2005},
  pages     = {675-682},
  url       = {https://mlanthology.org/neurips/2005/kreidl2005neurips-inference/}
}