论文标题
在有效实现低率矩阵优化的矩阵启用梯度算法上
On the Efficient Implementation of the Matrix Exponentiated Gradient Algorithm for Low-Rank Matrix Optimization
论文作者
论文摘要
频谱上的凸优化,即具有单位跟踪的所有实际$ n \ times n $阳性半矩阵的集合,在机器学习,信号处理和统计数据中具有重要的应用,主要是用于低级矩阵优化问题的凸放松。在凸优化的一阶方法理论中,它也是最突出的例子之一,在该理论中,非欧几里得方法可以比其欧几里得对应物显着可取。特别是,理想的选择是基于(负)von neumann熵引起的布雷格曼距离的矩阵启用梯度(MEG)方法。不幸的是,实施MEG需要对每次迭代进行完整的SVD计算,这对于高维问题不可扩展。在这项工作中,我们提出了具有确定性和随机梯度的MEG的有效实现,这些梯度是针对低级矩阵进行优化的,并且仅在每次迭代中使用单个低级别的SVD计算。我们还提供有效汇总的证书,以正确地融合我们的方法。主要是,我们证明在严格的互补条件下,建议的方法从``温暖启动''初始化汇聚,其速度与基于全SVD的同类产品相似。最后,我们带来了经验实验,这些实验既支持我们的理论发现,又证明了我们方法的实际吸引力。
Convex optimization over the spectrahedron, i.e., the set of all real $n\times n$ positive semidefinite matrices with unit trace, has important applications in machine learning, signal processing and statistics, mainly as a convex relaxation for optimization problems with low-rank matrices. It is also one of the most prominent examples in the theory of first-order methods for convex optimization in which non-Euclidean methods can be significantly preferable to their Euclidean counterparts. In particular, the desirable choice is the Matrix Exponentiated Gradient (MEG) method which is based on the Bregman distance induced by the (negative) von Neumann entropy. Unfortunately, implementing MEG requires a full SVD computation on each iteration, which is not scalable to high-dimensional problems. In this work we propose an efficient implementations of MEG, both with deterministic and stochastic gradients, which are tailored for optimization with low-rank matrices, and only use a single low-rank SVD computation on each iteration. We also provide efficiently-computable certificates for the correct convergence of our methods. Mainly, we prove that under a strict complementarity condition, the suggested methods converge from a ``warm-start" initialization with similar rates to their full-SVD-based counterparts. Finally, we bring empirical experiments which both support our theoretical findings and demonstrate the practical appeal of our methods.