ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1302.2325
48
189

Conditional Gradient Algorithms for Norm-Regularized Smooth Convex Optimization

10 February 2013
Zaïd Harchaoui
A. Juditsky
A. Nemirovski
ArXivPDFHTML
Abstract

Motivated by some applications in signal processing and machine learning, we consider two convex optimization problems where, given a cone KKK, a norm ∥⋅∥\|\cdot\|∥⋅∥ and a smooth convex function fff, we want either 1) to minimize the norm over the intersection of the cone and a level set of fff, or 2) to minimize over the cone the sum of fff and a multiple of the norm. We focus on the case where (a) the dimension of the problem is too large to allow for interior point algorithms, (b) ∥⋅∥\|\cdot\|∥⋅∥ is "too complicated" to allow for computationally cheap Bregman projections required in the first-order proximal gradient algorithms. On the other hand, we assume that {it is relatively easy to minimize linear forms over the intersection of KKK and the unit ∥⋅∥\|\cdot\|∥⋅∥-ball}. Motivating examples are given by the nuclear norm with KKK being the entire space of matrices, or the positive semidefinite cone in the space of symmetric matrices, and the Total Variation norm on the space of 2D images. We discuss versions of the Conditional Gradient algorithm capable to handle our problems of interest, provide the related theoretical efficiency estimates and outline some applications.

View on arXiv
Comments on this paper