There's definitely a QR decomposition in there for me since solving A = X Y' for X is X = A Y (Y' * Y)^-1 and you need some means to compute the inverse of that (small) matrix.
On Tue, Jan 8, 2013 at 5:27 PM, Ted Dunning <[email protected]> wrote: > This particular part of the algorithm can be seen as similar to a least > squares problem that might normally be solved by QR. I don't think that > the updates are quite the same, however. > > On Tue, Jan 8, 2013 at 3:10 PM, Sebastian Schelter <[email protected]> wrote: > >> This factorization is iteratively refined. In each iteration, ALS first >> fixes the item-feature vectors and solves a least-squares problem for >> each user and then fixes the user-feature vectors and solves a >> least-squares problem for each item. >>
