Asymmetry helps: eigenvalue and eigenvector analyses of asymmetrically perturbed low-rank matrices

ANNALS OF STATISTICS(2021)

引用 38|浏览61
暂无评分
摘要
This paper is concerned with the interplay between statistical asymmetry and spectral methods. Suppose we are interested in estimating a rank-1 and symmetric matrix M* is an element of R-n(xn), yet only a randomly perturbed version M is observed. The noise matrix M - M* is composed of independent (but not necessarily homoscedastic) entries and is, therefore, not symmetric in general. This might arise if, for example, when we have two independent samples for each entry of M* and arrange them in an asymmetric fashion. The aim is to estimate the leading eigenvalue and the leading eigenvector of M*. We demonstrate that the leading eigenvalue of the data matrix M can be O (root n) times more accurate (up to some log factor) than its (unadjusted) leading singular value of M in eigenvalue estimation. Moreover, the eigendecomposition approach is fully adaptive to heteroscedasticity of noise, without the need of any prior knowledge about the noise distributions. In a nutshell, this curious phenomenon arises since the statistical asymmetry automatically mitigates the bias of the eigenvalue approach, thus eliminating the need of careful bias correction. Additionally, we develop appealing nonasymptotic eigenvector perturbation bounds; in particular, we are able to bound the perturbation of any linear function of the leading eigenvector of M (e.g., entrywise eigenvector perturbation). We also provide partial theory for the more general rank-r case. The takeaway message is this: arranging the data samples in an asymmetric manner and performing eigendecomposition could sometimes be quite beneficial.
更多
查看译文
关键词
Spectral methods,eigenvalue perturbation,entrywise eigenvector perturbation,linear form of eigenvectors,heteroscedasticity
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要