Efficient Data Representations That Preserve Information
Abstract
A fundamental issue in computational learning theory, as well as in biological information processing, is the best possible relationship between model representation complexity and its prediction accuracy. Clearly, we expect more complex models that require longer data representation to be more accurate. Can one provide a quantitative, yet general, formulation of this trade-off? In this talk I will discuss this question from Shannon’s Information Theory perspective. I will argue that this trade-off can be traced back to the basic duality between source and channel coding and is also related to the notion of “coding with side information”. I will review some of the theoretical achievability results for such relevant data representations and discuss our algorithms for extracting them. I will then demonstrate the application of these ideas for the analysis of natural language corpora and speculate on possibly-universal aspects of human language that they reveal. Based on joint works with Ran Bacharach, Gal Chechik, Amir Globerson, Amir Navot, and Noam Slonim.
Cite
Text
Tishby. "Efficient Data Representations That Preserve Information." International Conference on Algorithmic Learning Theory, 2003. doi:10.1007/978-3-540-39624-6_4Markdown
[Tishby. "Efficient Data Representations That Preserve Information." International Conference on Algorithmic Learning Theory, 2003.](https://mlanthology.org/alt/2003/tishby2003alt-efficient/) doi:10.1007/978-3-540-39624-6_4BibTeX
@inproceedings{tishby2003alt-efficient,
title = {{Efficient Data Representations That Preserve Information}},
author = {Tishby, Naftali},
booktitle = {International Conference on Algorithmic Learning Theory},
year = {2003},
pages = {16},
doi = {10.1007/978-3-540-39624-6_4},
url = {https://mlanthology.org/alt/2003/tishby2003alt-efficient/}
}