Statistical Tests for Optimization Efficiency - Robotics Institute Carnegie Mellon University

Statistical Tests for Optimization Efficiency

Levi Boyles, Anoop Korattikara, Deva Ramanan, and Max Welling
Conference Paper, Proceedings of (NeurIPS) Neural Information Processing Systems, pp. 2196 - 2204, December, 2011

Abstract

Learning problems, such as logistic regression, are typically formulated as pure optimization problems defined on some loss function. We argue that this view ignores the fact that the loss function depends on stochastically generated data which in turn determines an intrinsic scale of precision for statistical estimation. By considering the statistical properties of the update variables used during the optimization (e.g. gradients), we can construct frequentist hypothesis tests to determine the reliability of these updates. We utilize subsets of the data for com-puting updates, and use the hypothesis tests for determining when the batch-size needs to be increased. This provides computational benefits and avoids overfitting by stopping when the batch-size has become equal to size of the full dataset. Moreover, the proposed algorithms depend on a single interpretable parameter – the probability for an update to be in the wrong direction – which is set to a single value across all algorithms and datasets. In this paper, we illustrate these ideas on three L 1 regularized coordinate descent algorithms: L 1 -regularized L 2 -loss SVMs, L 1 -regularized logistic regression, and the Lasso, but we emphasize that the underlying methods are much more generally applicable.

BibTeX

@conference{Boyles-2011-121211,
author = {Levi Boyles and Anoop Korattikara and Deva Ramanan and Max Welling},
title = {Statistical Tests for Optimization Efficiency},
booktitle = {Proceedings of (NeurIPS) Neural Information Processing Systems},
year = {2011},
month = {December},
pages = {2196 - 2204},
}