Early stopping and non-parametric regression: An optimal data-dependent stopping rule

06/15/2013
by   Garvesh Raskutti, et al.
0

The strategy of early stopping is a regularization technique based on choosing a stopping time for an iterative algorithm. Focusing on non-parametric regression in a reproducing kernel Hilbert space, we analyze the early stopping strategy for a form of gradient-descent applied to the least-squares loss function. We propose a data-dependent stopping rule that does not involve hold-out or cross-validation data, and we prove upper bounds on the squared error of the resulting function estimate, measured in either the L^2(P) and L^2(P_n) norm. These upper bounds lead to minimax-optimal rates for various kernel classes, including Sobolev smoothness classes and other forms of reproducing kernel Hilbert spaces. We show through simulation that our stopping rule compares favorably to two other stopping rules, one based on hold-out data and the other based on Stein's unbiased risk estimate. We also establish a tight connection between our early stopping strategy and the solution path of a kernel ridge regression estimator.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/25/2018

Early Stopping for Nonparametric Testing

Early stopping of iterative algorithms is an algorithmic regularization ...
research
07/14/2020

Early stopping and polynomial smoothing in regression with reproducing kernels

In this paper we study the problem of early stopping for iterative learn...
research
01/23/2018

Non-parametric sparse additive auto-regressive network models

Consider a multi-variate time series (X_t)_t=0^T where X_t ∈R^d which ma...
research
10/19/2015

NYTRO: When Subsampling Meets Early Stopping

Early stopping is a well known approach to reduce the time complexity fo...
research
01/27/2023

Conformal inference is (almost) free for neural networks trained with early stopping

Early stopping based on hold-out data is a popular regularization techni...
research
08/20/2020

Minimum discrepancy principle strategy for choosing k in k-NN regression

This paper presents a novel data-driven strategy to choose the hyperpara...
research
09/22/2020

Risk upper bounds for RKHS ridge group sparse estimator in the regression model with non-Gaussian and non-bounded error

We consider the problem of estimating a meta-model of an unknown regress...

Please sign up or login with your details

Forgot password? Click here to reset