Evaluating Probabilistic Reasoning Systems

Abstract

Over the past two decades a variety of exact and approximate algorithms were developed across several communities (e.g. UAI, NIPS, SAT/CSPs) for answering optimization and likelihood queries over probabilistic graphical models. Since all these tasks are NP-hard, theoretical guarantees are rare and empirical evaluation becomes a central evaluation tool. Yet, the empirical comparison between algorithms requires agreement on representations, benchmarks and evaluation criteria which is challenging, especially in the context of approximation algorithms. Some communities have already addressed similar challenges through yearly empirical evaluations and competitions (e.g. SAT, CSP and planning) which proved effective, leading to algorithmic advances and to software development and dissemination. We believe that such an effort could benefit probabilistic inference algorithms as well. Probabilistic reasoning presents additional challenges, however, as it tends to be harder, requires heterogenous knowledge representation frameworks, and must deal with the issue of evaluating approximate inference algorithms.

Cite

Text

Darwiche and Dechter. "Evaluating Probabilistic Reasoning Systems." Conference on Uncertainty in Artificial Intelligence, 2008.

Markdown

[Darwiche and Dechter. "Evaluating Probabilistic Reasoning Systems." Conference on Uncertainty in Artificial Intelligence, 2008.](https://mlanthology.org/uai/2008/darwiche2008uai-evaluating/)

BibTeX

@inproceedings{darwiche2008uai-evaluating,
  title     = {{Evaluating Probabilistic Reasoning Systems}},
  author    = {Darwiche, Adnan and Dechter, Rina},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2008},
  url       = {https://mlanthology.org/uai/2008/darwiche2008uai-evaluating/}
}