Examples of pathological dynamics of the subgradient method for Lipschitz path-differentiable functions
We show that the vanishing stepsize subgradient method – widely adopted for machine learning applications – can display rather messy behavior even in the presence of favorable assumptions. We establish that convergence of bounded subgradient sequences may fail even with a Whitney stratifiable objective function satisfying the Kurdyka-Lojasiewicz inequality. Moreover, when the objective function is path-differentiable we show that various properties all may fail to occur: criticality of the limit points, convergence of the sequence, convergence in values, codimension one of the accumulation set, equality of the accumulation and essential accumulation sets, connectedness of the essential accumulation set, spontaneous slowdown, oscillation compensation, and oscillation perpendicularity to the accumulation set.
READ FULL TEXT