Skip to content
On this page

On Optimisation in Matrix Completion

Statisticians get their first taste of optimisation when they learn about penalised linear regression:

minβYXβ22+λβp.\min_{\beta} \|Y - X \beta\|_2^2 + \lambda\|\beta\|_p.

It turns out that there's an equivalent formulation (as a constrained minimisation) that provides a little more intuition:

minYXβ22 s.t. βps,\min \|Y - X \beta\|_2^2 \text{ s.t. } \|\beta\|_p \leq s,
The OG image.
The OG image.

where there is a one-to-one relation between λ\lambda and ss. This different view provides a geometric interpretation: depending on the geometry induced by the lpl_p ball that is the constraint set, you're going to get different kinds of solutions.

Note that the above equivalence is exact only when the loss and the penalty are both convex. In the matrix completion setting ([[penalty-project]]), this is no longer the case.

The (classical) convex program (Candes & Recht, 2009)Candes:2009kj is given by:

minW s.t. A(W)y22ϵ,\min \|W\|_\star \text{ s.t. } \|\mathcal{A}(W) - y \|_2^2 \leq \epsilon,

where setting ϵ=0\epsilon = 0 reduces to the noise-less setting. The more recent development using a penalised linear model is given by:

minWA(W)y22+λW\min_{W} \| \mathcal{A}(W) - y \|_2^2 + \lambda \|W\|_{\star}

Since they're both convex, you have equivalence just like for the Lasso.1

Non-Convex

What happens when we move to non-convex penalties though? Then, I think what you have is a duality gap (I could be wrong). If we use our penalty WWF\frac{\|{W}_{\star}}{\|{W}_{F}}, then these two views are no longer equivalent. And here I think actually going back to the penalised loss function gives better results. For one thing, it's no longer a convex program, amenable to simple solvers.


  1. Candes, E.J. & Recht, B., 2009. Exact matrix completion via convex optimization. Foundations of Computational Mathematics, 9(6), pp.717–772.
  2. I'm pretty sure that people have profitably used this equivalence in this context.
Edit this page
Last updated on 1/30/2022