美文网首页
matrix decomposition

matrix decomposition

作者: balding | 来源:发表于2018-12-01 21:47 被阅读0次

    1. CLUSTERED MATRIX APPROXIMATION

    SIAM [Society for Industrial and Applied Mathematics] 2016

    1. develop a clustered matrix approximation framework
      1. Divide the matrix into different groups
    2. particularly well suited for problems with large-scale sparse matrices

    Abstract

    benefits
    1. preserve the important structure presented in the original matrix;
    2. contain both global-scale and local-scale information
    3. efficient both in computational speed and memory usage
    4. resulting approximations more accurate with less memory usage than truncated SVD approximations, which are optimal with respect to rank.
    5. flexible: modified in various ways [what ways?]
    method
    1. derive a probabilistic approach that uses randomness to compute a clustered matrix approximation within the developed framework?.
    2. We further prove deterministic and probabilistic bounds of the resulting approximation error.

    Introduction

    • approximate A \in R ^{m \times n} with low-rank \tilde{A}, \min _{rank(\tilde{A})=k} ||A-\tilde{A}||_F has been studied extensively in the literature
    • bottleneck: with large amounts of data is often memory consumption and not the rank of the approximation
      • In this setting, our method produces a considerably more accurate matrix approximation with memory usage that is the same as or less than that of the truncated SVD approximation, which is optimal with respect to rank.
    steps
    1. a clustering or coclustering step so that the rows and columns of a given matrix are partitioned into a number of groups;
    2. reordering the rows and columns according to cluster a�affiliation and extracting sufficiently dense blocks;
    3. computing low-rank approximations of these dense blocks;
    4. combining the blockwise approximations into an approximation of the entire
      matrix.
    explanation

    Computing truncated SVD approximations is relatively expensive, one may wish to trade off� computation time against some slight loss inaccuracy. Probabilistic algorithms for matrix approximation [21] give the means for this kind of trade-off�.

    • develop and detail the method
    Preliminaries

    rely on a number of concepts:

    • efficient graph clustering;
    • bipartite graph coclustering;
    • low-rank matrix approximations;
    • stochastic methods for low rank matrix approximations.

    Bipartite graph: is a graph whose vertices can be divided into two disjoint and independent sets U and V such that every edge connects a vertex in U to one in V . U and V are usually called the parts of the graph

    1. assume that we can partition the graph and obtain c disjoint sets

      \cal{V_1}, \dots, V_c , with m_i = |\cal{V_{i}}| . Without loss of generality, we can assume that vertices in \cal{V_1}, \dots, V_c are sorted in strictly increasing order. Then matrix has form
      A= \begin{bmatrix} A_{11} & \dots & A_{1c}\\ \vdots & \ddots & \vdots\\ A_{c1} & \dots & A_{cc} \end{bmatrix}

    1. given A and rank k
    2. standard Gausion matrix \Omega \in R^{n \times (k+p)} , p is a small oversampling parameter (typically set to 5 ~ 10)
    3. Y= A \Omega
    4. Q = orth (Y) : orthonormal basis
    5. the corresponding low rank approximation is given by A \approx \tilde{A} = QQ^TA
    6. Q^T A = \bar{W} \bar{\Sigma} \bar{V}^T
    7. \hat{A} = (Q \bar{W}) \bar{\Sigma} \bar{V}^T = \bar {U}\bar{\Sigma} \bar{V}^T

    Clustered low rank matrix approximations

    However , it used existing clusters method... as it is more about computing U_i,V_i,S_{ij}

    End.

    2. Improved Nystrom Low-Rank Approximation and Error Analysis

    This paper presents detailed studies on the Nystrom sampling scheme and in particular, an error analysis that directly relates the Nystrom approximation quality with the encoding powers of the landmark points.

    Introduction

    Kernel methods play a central role in machine learning and have demonstrated huge success in modeling real-world data with highly complex, nonlinear structures.

    相关文章

      网友评论

          本文标题:matrix decomposition

          本文链接:https://www.haomeiwen.com/subject/qcfdcqtx.html