Learning Graph Laplacian with MCP

10/22/2020
by   Yangjing Zhang, et al.
0

Motivated by the observation that the ability of the ℓ_1 norm in promoting sparsity in graphical models with Laplacian constraints is much weakened, this paper proposes to learn graph Laplacian with a non-convex penalty: minimax concave penalty (MCP). For solving the MCP penalized graphical model, we design an inexact proximal difference-of-convex algorithm (DCA) and prove its convergence to critical points. We note that each subproblem of the proximal DCA enjoys the nice property that the objective function in its dual problem is continuously differentiable with a semismooth gradient. Therefore, we apply an efficient semismooth Newton method to subproblems of the proximal DCA. Numerical experiments on various synthetic and real data sets demonstrate the effectiveness of the non-convex penalty MCP in promoting sparsity. Compared with the state-of-the-art method <cit.>, our method is demonstrated to be more efficient and reliable for learning graph Laplacian with MCP.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset