55
3

Optimal Reduced Isotonic Regression

Abstract

Isotonic regression is a shape-constrained nonparametric regression in which the regression is an increasing step function. For nn data points, the number of steps in the isotonic regression may be as large as nn. As a result, standard isotonic regression has been criticized as overfitting the data or making the representation too complicated. So-called "reduced" isotonic regression constrains the outcome to be a specified number of steps bb, bnb \leq n. However, because the previous algorithms for finding the reduced L2L_2 regression took Θ(n+bm2)\Theta(n+bm^2) time, where mm is the number of steps of the unconstrained isotonic regression, researchers felt that the algorithms were too slow and instead used approximations. Other researchers had results that were approximations because they used a greedy top-down approach. Here we give an algorithm to find an exact solution in Θ(n+bm)\Theta(n+bm) time, and a simpler algorithm taking Θ(n+bmlogm)\Theta(n+b m \log m) time. These algorithms also determine optimal kk-means clustering of weighted 1-dimensional data.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.