Outer product models are the simplest matrix model that couples entries in each row and each column. Given a matrix , we would like to approximate it with the outer product of two vectors and , such that the squared error is minimized:
where
We note that is a rank 1 matrix.
Gradients
If is fully observed, the objective becomes
Taking partial derivatives with respect to and and letting , we have
derivation
Using the chain rule,
The derivative of the Frobenius norm is
Also,
where
Therefore,
Similarly, we have
Convexity
Given the gradients, the hessian matrix is simply calculated
The Hessian matrix at the origin is
which is not positive semi-definite, unless is a zero matrix. Therefore the problem is non-convex for all dimensions .
Solving with Lagrange Multiplier
Using the property that
we can rewrite the objective, using the properties of trace (invariance under transpose and circular shift):
which implies should be proportional to the principal eigenvector of the matrix , and should be proportional to the principal eigenvector of the matrix . This can be simply done by SVD.