Isotonic regression in general dimensions

From MaRDI portal
Publication:2328048

DOI10.1214/18-AOS1753zbMATH Open1437.62124arXiv1708.09468MaRDI QIDQ2328048FDOQ2328048


Authors: Qiyang Han, Tengyao Wang, Sabyasachi Chatterjee, Richard Samworth Edit this on Wikidata


Publication date: 9 October 2019

Published in: The Annals of Statistics (Search for Journal in Brave)

Abstract: We study the least squares regression function estimator over the class of real-valued functions on [0,1]d that are increasing in each coordinate. For uniformly bounded signals and with a fixed, cubic lattice design, we establish that the estimator achieves the minimax rate of order nmin2/(d+2),1/d in the empirical L2 loss, up to poly-logarithmic factors. Further, we prove a sharp oracle inequality, which reveals in particular that when the true regression function is piecewise constant on k hyperrectangles, the least squares estimator enjoys a faster, adaptive rate of convergence of (k/n)min(1,2/d), again up to poly-logarithmic factors. Previous results are confined to the case dleq2. Finally, we establish corresponding bounds (which are new even in the case d=2) in the more challenging random design setting. There are two surprising features of these results: first, they demonstrate that it is possible for a global empirical risk minimisation procedure to be rate optimal up to poly-logarithmic factors even when the corresponding entropy integral for the function class diverges rapidly; second, they indicate that the adaptation rate for shape-constrained estimators can be strictly worse than the parametric rate.


Full work available at URL: https://arxiv.org/abs/1708.09468




Recommendations




Cites Work


Cited In (51)

Uses Software





This page was built for publication: Isotonic regression in general dimensions

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2328048)