arXiv Analytics

Sign in

arXiv:1802.02163 [stat.ML]AbstractReferencesReviewsResources

How to Make Causal Inferences Using Texts

Naoki Egami, Christian J. Fong, Justin Grimmer, Margaret E. Roberts, Brandon M. Stewart

Published 2018-02-06Version 1

New text as data techniques offer a great promise: the ability to inductively discover measures that are useful for testing social science theories of interest from large collections of text. We introduce a conceptual framework for making causal inferences with discovered measures as a treatment or outcome. Our framework enables researchers to discover high-dimensional textual interventions and estimate the ways that observed treatments affect text-based outcomes. We argue that nearly all text-based causal inferences depend upon a latent representation of the text and we provide a framework to learn the latent representation. But estimating this latent representation, we show, creates new risks: we may introduce an identification problem or overfit. To address these risks we describe a split-sample framework and apply it to estimate causal effects from an experiment on immigration attitudes and a study on bureaucratic response. Our work provides a rigorous foundation for text-based causal inferences.

Related articles: Most relevant | Search more
arXiv:1907.09708 [stat.ML] (Published 2019-07-23)
Node Attribute Generation on Graphs
arXiv:1102.1492 [stat.ML] (Published 2011-02-08, updated 2011-10-26)
On Nonparametric Guidance for Learning Autoencoder Representations
arXiv:2007.07142 [stat.ML] (Published 2020-07-14)
Extendable and invertible manifold learning with geometry regularized autoencoders