{ "id": "2202.00622", "version": "v1", "published": "2022-02-01T18:15:24.000Z", "updated": "2022-02-01T18:15:24.000Z", "title": "Datamodels: Predicting Predictions from Training Data", "authors": [ "Andrew Ilyas", "Sung Min Park", "Logan Engstrom", "Guillaume Leclerc", "Aleksander Madry" ], "categories": [ "stat.ML", "cs.CV", "cs.LG" ], "abstract": "We present a conceptual framework, datamodeling, for analyzing the behavior of a model class in terms of the training data. For any fixed \"target\" example $x$, training set $S$, and learning algorithm, a datamodel is a parameterized function $2^S \\to \\mathbb{R}$ that for any subset of $S' \\subset S$ -- using only information about which examples of $S$ are contained in $S'$ -- predicts the outcome of training a model on $S'$ and evaluating on $x$. Despite the potential complexity of the underlying process being approximated (e.g., end-to-end training and evaluation of deep neural networks), we show that even simple linear datamodels can successfully predict model outputs. We then demonstrate that datamodels give rise to a variety of applications, such as: accurately predicting the effect of dataset counterfactuals; identifying brittle predictions; finding semantically similar examples; quantifying train-test leakage; and embedding data into a well-behaved and feature-rich representation space. Data for this paper (including pre-computed datamodels as well as raw predictions from four million trained deep neural networks) is available at https://github.com/MadryLab/datamodels-data .", "revisions": [ { "version": "v1", "updated": "2022-02-01T18:15:24.000Z" } ], "analyses": { "keywords": [ "training data", "predicting predictions", "million trained deep neural networks", "simple linear datamodels", "feature-rich representation space" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }