Poster
Efficient Alternating Minimization with Applications to Weighted Low Rank Approximation
Zhao Song · Mingquan Ye · Junze Yin · Lichen Zhang
Hall 3 + Hall 2B #342
[
Abstract
]
Sat 26 Apr midnight PDT
— 2:30 a.m. PDT
Abstract:
Weighted low rank approximation is a fundamental problem in numerical linear algebra, and it has many applications in machine learning. Given a matrix M∈Rn×n, a non-negative weight matrix W∈Rn×n≥0, a parameter k, the goal is to output two matrices X,Y∈Rn×k such that |W∘(M−XY⊤)|F is minimized, where ∘ denotes the Hadamard product. It naturally generalizes the well-studied low rank matrix completion problem. Such a problem is known to be NP-hard and even hard to approximate assuming the Exponential Time Hypothesis. Meanwhile, alternating minimization is a good heuristic solution for weighted low rank approximation. In particular, [Li, Liang and Risteski, ICML'16] shows that, under mild assumptions, alternating minimization does provide provable guarantees. In this work, we develop an efficient and robust framework for alternating minimization that allows the alternating updates to be computed approximately. For weighted low rank approximation, this improves the runtime of [Li, Liang and Risteski, ICML'16] from |W|0k2 to |W|0k where |W|0 denotes the number of nonzero entries of the weight matrix. At the heart of our framework is a high-accuracy multiple response regression solver together with a robust analysis of alternating minimization.
Live content is unavailable. Log in and register to view live content