1

I am trying to implement Nesterov's first method to solve convex piece-wise linear optimization problem, from this website:

https://blogs.princeton.edu/imabandit/2013/04/01/acceleratedgradientdescent/

But then, such $\beta$ does not exist convex piece-wise linear function. So I am wondering what shall I put into $\beta$ for my implementations.

PS: LP is not feasible because there are $2^{80}$ such hyper planes.

nicoguaro
  • 8,500
  • 6
  • 23
  • 49
user40780
  • 243
  • 1
  • 9

1 Answers1

6

A piecewise linear function is not differentiable (except in the trivial case), so as you noticed this method cannot be applied - the gradient does not exist, let alone its Lipschitz constant beta.

If you want to use a variant of Nesterov's accelerated algorithm, you have two options:

  1. You replace your function by a smooth approximation and apply an accelerated gradient descent; this is described in Nesterov's paper Smooth minimization of non-smooth functions, Mathematical Programming May 2005, Volume 103, Issue 1, pp 127-152, or

  2. you use his accelerated subgradient scheme for nonsmooth convex functions; this is described in his paper Primal-dual subgradient methods for convex problems, Mathematical Programming August 2009, Volume 120, Issue 1, pp 221-259.

Christian Clason
  • 12,301
  • 3
  • 48
  • 68