{ "id": "1705.00612", "version": "v1", "published": "2017-04-29T21:35:26.000Z", "updated": "2017-04-29T21:35:26.000Z", "title": "Thermodynamic cost and benefit of data representations", "authors": [ "Susanne Still" ], "categories": [ "cond-mat.stat-mech" ], "abstract": "This paper takes a thermodynamic approach to addressing the problem of how to represent data efficiently and meaningfully, a problem at the heart of learning and adaptation in both biological and artificial systems. Thermodynamic analysis of an information engine's cyclic operation reveals information theoretic quantities that are setting limits on performance. If run at fixed temperature, dissipation is lower bounded by a term proportional to irrelevant information. Data representation strategies that are optimal in the sense of minimizing dissipation must therefore strive to retain only relevant information. When an information engine is allowed to make use of a temperature difference, it can produce net work output, for which an upper bound is derived. Maximizing the bound yields a direct derivation of the Information Bottleneck method, a known technique in signal processing and machine learning, used precisely to filter relevant information from irrelevant clutter.", "revisions": [ { "version": "v1", "updated": "2017-04-29T21:35:26.000Z" } ], "analyses": { "keywords": [ "thermodynamic cost", "produce net work output", "information engines cyclic operation", "data representation strategies", "information theoretic quantities" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }