73
16

Generic Coreset for Scalable Learning of Monotonic Kernels: Logistic Regression, Sigmoid and more

Abstract

Coreset (or core-set) is a small weighted \emph{subset} QQ of an input set PP with respect to a given \emph{monotonic} function f:RRf:\mathbb{R}\to\mathbb{R} that \emph{provably} approximates its fitting loss pPf(px)\sum_{p\in P}f(p\cdot x) to \emph{any} given xRdx\in\mathbb{R}^d. Using QQ we can obtain approximation of xx^* that minimizes this loss, by running \emph{existing} optimization algorithms on QQ. In this work we provide: (i) A lower bound which proves that there are sets with no coresets smaller than n=Pn=|P| for general monotonic loss functions. (ii) A proof that, under a natural assumption that holds e.g. for logistic regression and the sigmoid activation functions, a small coreset exists for \emph{any} input PP. (iii) A generic coreset construction algorithm that computes such a small coreset QQ in O(nd+nlogn)O(nd+n\log n) time, and (iv) Experimental results which demonstrate that our coresets are effective and are much smaller in practice than predicted in theory.

View on arXiv
Comments on this paper