SVM via Saddle Point Optimization: New Bounds and Distributed Algorithms

We study two important SVM variants: hard-margin SVM (for linearly separable cases) and -SVM (for linearly non-separable cases). We propose new algorithms from the perspective of saddle point optimization. Our algorithms achieve -approximations with running time for both variants, where is the number of points and is the dimensionality. To the best of our knowledge, the current best algorithm for -SVM is based on quadratic programming approach which requires time in worst case~\cite{joachims1998making,platt199912}. In the paper, we provide the first nearly linear time algorithm for -SVM. The current best algorithm for hard margin SVM achieved by Gilbert algorithm~\cite{gartner2009coresets} requires time. Our algorithm improves the running time by a factor of . Moreover, our algorithms can be implemented in the distributed settings naturally. We prove that our algorithms require communication cost, where is the number of clients, which almost matches the theoretical lower bound. Numerical experiments support our theory and show that our algorithms converge faster on high dimensional, large and dense data sets, as compared to previous methods.
View on arXiv