arXiv Analytics

Sign in

arXiv:2109.03795 [stat.ML]AbstractReferencesReviewsResources

Desiderata for Representation Learning: A Causal Perspective

Yixin Wang, Michael I. Jordan

Published 2021-09-08Version 1

Representation learning constructs low-dimensional representations to summarize essential features of high-dimensional data. This learning problem is often approached by describing various desiderata associated with learned representations; e.g., that they be non-spurious, efficient, or disentangled. It can be challenging, however, to turn these intuitive desiderata into formal criteria that can be measured and enhanced based on observed data. In this paper, we take a causal perspective on representation learning, formalizing non-spuriousness and efficiency (in supervised representation learning) and disentanglement (in unsupervised representation learning) using counterfactual quantities and observable consequences of causal assertions. This yields computable metrics that can be used to assess the degree to which representations satisfy the desiderata of interest and learn non-spurious and disentangled representations from single observational datasets.

Related articles: Most relevant | Search more
arXiv:2312.09061 [stat.ML] (Published 2023-12-14)
Fair Clustering: A Causal Perspective
arXiv:2202.06844 [stat.ML] (Published 2022-02-14)
On Pitfalls of Identifiability in Unsupervised Learning. A Note on: "Desiderata for Representation Learning: A Causal Perspective"
arXiv:2107.07853 [stat.ML] (Published 2021-07-16)
A Causal Perspective on Meaningful and Robust Algorithmic Recourse