arXiv Analytics

Sign in

arXiv:1806.10701 [stat.ML]AbstractReferencesReviewsResources

Empirical Risk Minimization and Stochastic Gradient Descent for Relational Data

Victor Veitch, Morgane Austern, Wenda Zhou, David M. Blei, Peter Orbanz

Published 2018-06-27Version 1

Empirical risk minimization is the principal tool for prediction problems, but its extension to relational data remains unsolved. We solve this problem using recent advances in graph sampling theory. We (i) define an empirical risk for relational data and (ii) obtain stochastic gradients for this risk that are automatically unbiased. The key ingredient is to consider the method by which data is sampled from a graph as an explicit component of model design. Theoretical results establish that the choice of sampling scheme is critical. By integrating fast implementations of graph sampling schemes with standard automatic differentiation tools, we are able to solve the risk minimization in a plug-and-play fashion even on large datasets. We demonstrate empirically that relational ERM models achieve state-of-the-art results on semi-supervised node classification tasks. The experiments also confirm the importance of the choice of sampling scheme.

Related articles: Most relevant | Search more
arXiv:1806.01796 [stat.ML] (Published 2018-06-05)
Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate
arXiv:2212.14883 [stat.ML] (Published 2022-12-30)
Online Statistical Inference for Contextual Bandits via Stochastic Gradient Descent
arXiv:1510.04822 [stat.ML] (Published 2015-10-16)
SGD with Variance Reduction beyond Empirical Risk Minimization