Generalized singular value decomposition

Last updated

In linear algebra, the generalized singular value decomposition (GSVD) is the name of two different techniques based on the singular value decomposition (SVD). The two versions differ because one version decomposes two matrices (somewhat like the higher-order or tensor SVD) and the other version uses a set of constraints imposed on the left and right singular vectors of a single-matrix SVD.

Contents

First version: two-matrix decomposition

The generalized singular value decomposition (GSVD) is a matrix decomposition on a pair of matrices which generalizes the singular value decomposition. It was introduced by Van Loan [1] in 1976 and later developed by Paige and Saunders, [2] which is the version described here. In contrast to the SVD, the GSVD decomposes simultaneously a pair of matrices with the same number of columns. The SVD and the GSVD, as well as some other possible generalizations of the SVD, [3] [4] [5] are extensively used in the study of the conditioning and regularization of linear systems with respect to quadratic semi-norms. In the following, let , or .

Definition

The generalized singular value decomposition of matrices and is

where

We denote , , , and . While is diagonal, is not always diagonal, because of the leading rectangular zero matrix; instead is "bottom-right-diagonal".

Variations

There are many variations of the GSVD. These variations are related to the fact that it is always possible to multiply from the left by where is an arbitrary unitary matrix. We denote

Here are some variations of the GSVD:

Generalized singular values

A generalized singular value of and is a pair such that

We have


By these properties we can show that the generalized singular values are exactly the pairs . We have

Therefore

This expression is zero exactly when and for some .

In, [2] the generalized singular values are claimed to be those which solve . However, this claim only holds when , since otherwise the determinant is zero for every pair ; this can be seen by substituting above.

Generalized inverse

Define for any invertible matrix , for any zero matrix , and for any block-diagonal matrix. Then define

It can be shown that as defined here is a generalized inverse of ; in particular a -inverse of . Since it does not in general satisfy , this is not the Moore–Penrose inverse; otherwise we could derive for any choice of matrices, which only holds for certain class of matrices.

Suppose , where and . This generalized inverse has the following properties:

Quotient SVD

A generalized singular ratio of and is . By the above properties, . Note that is diagonal, and that, ignoring the leading zeros, contains the singular ratios in decreasing order. If is invertible, then has no leading zeros, and the generalized singular ratios are the singular values, and and are the matrices of singular vectors, of the matrix . In fact, computing the SVD of is one of the motivations for the GSVD, as "forming and finding its SVD can lead to unnecessary and large numerical errors when is ill-conditioned for solution of equations". [2] Hence the sometimes used name "quotient SVD", although this is not the only reason for using GSVD. If is not invertible, then is still the SVD of if we relax the requirement of having the singular values in decreasing order. Alternatively, a decreasing order SVD can be found by moving the leading zeros to the back: , where and are appropriate permutation matrices. Since rank equals the number of non-zero singular values, .

Construction

Let

Then

We also have

Therefore

Since has orthonormal columns, . Therefore

We also have for each such that that

Therefore , and

Applications

The tensor GSVD is one of the comparative spectral decompositions, multi-tensor generalizations of the SVD, invented to simultaneously identify the similar and dissimilar among, and create a single coherent model from any data types, of any number and dimensions. Tensor Generalized Singular Value Decomposition following et int. Alter PLoS One 2015 and Alter NCI Physical Sciences in Oncology 2015.jpg
The tensor GSVD is one of the comparative spectral decompositions, multi-tensor generalizations of the SVD, invented to simultaneously identify the similar and dissimilar among, and create a single coherent model from any data types, of any number and dimensions.

The GSVD, formulated as a comparative spectral decomposition, [6] has been successfully applied to signal processing and data science, e.g., in genomic signal processing. [7] [8] [9]

These applications inspired several additional comparative spectral decompositions, i.e., the higher-order GSVD (HO GSVD) [10] and the tensor GSVD. [11] [12]

It has equally found applications to estimate the spectral decompositions of linear operators when the eigenfunctions are parameterized with a linear model, i.e. a reproducing kernel Hilbert space. [13]

Second version: weighted single-matrix decomposition

The weighted version of the generalized singular value decomposition (GSVD) is a constrained matrix decomposition with constraints imposed on the left and right singular vectors of the singular value decomposition. [14] [15] [16] This form of the GSVD is an extension of the SVD as such. Given the SVD of an m×n real or complex matrix M

where

Where I is the identity matrix and where and are orthonormal given their constraints ( and ). Additionally, and are positive definite matrices (often diagonal matrices of weights). This form of the GSVD is the core of certain techniques, such as generalized principal component analysis and Correspondence analysis.

The weighted form of the GSVD is called as such because, with the correct selection of weights, it generalizes many techniques (such as multidimensional scaling and linear discriminant analysis). [17]

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices which are Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

<span class="mw-page-title-main">Multivariate normal distribution</span> Generalization of the one-dimensional normal distribution to higher dimensions

In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution. Its importance derives mainly from the multivariate central limit theorem. The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables each of which clusters around a mean value.

<span class="mw-page-title-main">Singular value decomposition</span> Matrix decomposition

In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any matrix. It is related to the polar decomposition.

In mathematics, particularly in linear algebra, a skew-symmetricmatrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition

<span class="mw-page-title-main">Exterior algebra</span> Algebra of exterior/ wedge products

In mathematics, the exterior algebra, or Grassmann algebra, named after Hermann Grassmann, is an algebra that uses the exterior product or wedge product as its multiplication. In mathematics, the exterior product or wedge product of vectors is an algebraic construction used in geometry to study areas, volumes, and their higher-dimensional analogues. The exterior product of two vectors and  denoted by is called a bivector and lives in a space called the exterior square, a vector space that is distinct from the original space of vectors. The magnitude of can be interpreted as the area of the parallelogram with sides and  which in three dimensions can also be computed using the cross product of the two vectors. More generally, all parallel plane surfaces with the same orientation and area have the same bivector as a measure of their oriented area. Like the cross product, the exterior product is anticommutative, meaning that for all vectors and  but, unlike the cross product, the exterior product is associative.

In linear algebra, a QR decomposition, also known as a QR factorization or QU factorization, is a decomposition of a matrix A into a product A = QR of an orthonormal matrix Q and an upper triangular matrix R. QR decomposition is often used to solve the linear least squares problem and is the basis for a particular eigenvalue algorithm, the QR algorithm.

In mathematics, and in particular linear algebra, the Moore–Penrose inverse of a matrix is the most widely known generalization of the inverse matrix. It was independently described by E. H. Moore in 1920, Arne Bjerhammar in 1951, and Roger Penrose in 1955. Earlier, Erik Ivar Fredholm had introduced the concept of a pseudoinverse of integral operators in 1903. When referring to a matrix, the term pseudoinverse, without further specification, is often used to indicate the Moore–Penrose inverse. The term generalized inverse is sometimes used as a synonym for pseudoinverse.

In linear algebra and the theory of matrices, the Schur complement of a block matrix is defined as follows.

<span class="mw-page-title-main">Projection (linear algebra)</span> Idempotent linear transformation from a vector space to itself

In linear algebra and functional analysis, a projection is a linear transformation from a vector space to itself such that . That is, whenever is applied twice to any vector, it gives the same result as if it were applied once. It leaves its image unchanged. This definition of "projection" formalizes and generalizes the idea of graphical projection. One can also consider the effect of a projection on a geometrical object by examining the effect of the projection on points in the object.

In mathematics, the determinant of a skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depend on the size of the matrix. The value of this polynomial, when applied to the coefficients of a skew-symmetric matrix, is called the Pfaffian of that matrix. The term Pfaffian was introduced by Cayley (1852) who indirectly named them after Johann Friedrich Pfaff. The Pfaffian is nonvanishing only for 2n × 2n skew-symmetric matrices, in which case it is a polynomial of degree n.

Latent semantic analysis (LSA) is a technique in natural language processing, in particular distributional semantics, of analyzing relationships between a set of documents and the terms they contain by producing a set of concepts related to the documents and terms. LSA assumes that words that are close in meaning will occur in similar pieces of text. A matrix containing word counts per document is constructed from a large piece of text and a mathematical technique called singular value decomposition (SVD) is used to reduce the number of rows while preserving the similarity structure among columns. Documents are then compared by cosine similarity between any two columns. Values close to 1 represent very similar documents while values close to 0 represent very dissimilar documents.

<span class="mw-page-title-main">Singular value</span> Square roots of the eigenvalues of the self-adjoint operator

In mathematics, in particular functional analysis, the singular values, or s-numbers of a compact operator acting between Hilbert spaces and , are the square roots of the eigenvalues of the self-adjoint operator .

In mathematics, the Kronecker product, sometimes denoted by ⊗, is an operation on two matrices of arbitrary size resulting in a block matrix. It is a specialization of the tensor product from vectors to matrices and gives the matrix of the tensor product linear map with respect to a standard choice of basis. The Kronecker product is to be distinguished from the usual matrix multiplication, which is an entirely different operation. The Kronecker product is also sometimes called matrix direct product.

<span class="mw-page-title-main">Total least squares</span>

In applied statistics, total least squares is a type of errors-in-variables regression, a least squares data modeling technique in which observational errors on both dependent and independent variables are taken into account. It is a generalization of Deming regression and also of orthogonal regression, and can be applied to both linear and non-linear models.

In mathematics, the polar decomposition of a square real or complex matrix is a factorization of the form , where is a unitary matrix and is a positive semi-definite Hermitian matrix, both square and of the same size.

The Rayleigh–Ritz method is a direct numerical method of approximating eigenvalues, originated in the context of solving physical boundary value problems and named after Lord Rayleigh and Walther Ritz.

In mathematics, and in particular, algebra, a generalized inverse of an element x is an element y that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inverse of a matrix is to obtain a matrix that can serve as an inverse in some sense for a wider class of matrices than invertible matrices. Generalized inverses can be defined in any mathematical structure that involves associative multiplication, that is, in a semigroup. This article describes generalized inverses of a matrix .

In applied mathematics, Wahba's problem, first posed by Grace Wahba in 1965, seeks to find a rotation matrix between two coordinate systems from a set of (weighted) vector observations. Solutions to Wahba's problem are often used in satellite attitude determination utilising sensors such as magnetometers and multi-antenna GPS receivers. The cost function that Wahba's problem seeks to minimise is as follows:

In mathematics, low-rank approximation is a minimization problem, in which the cost function measures the fit between a given matrix and an approximating matrix, subject to a constraint that the approximating matrix has reduced rank. The problem is used for mathematical modeling and data compression. The rank constraint is related to a constraint on the complexity of a model that fits the data. In applications, often there are other constraints on the approximating matrix apart from the rank constraint, e.g., non-negativity and Hankel structure.

Generalized pencil-of-function method (GPOF), also known as matrix pencil method, is a signal processing technique for estimating a signal or extracting information with complex exponentials. Being similar to Prony and original pencil-of-function methods, it is generally preferred to those for its robustness and computational efficiency.

References

  1. Van Loan CF (1976). "Generalizing the Singular Value Decomposition". SIAM J. Numer. Anal. 13 (1): 76–83. Bibcode:1976SJNA...13...76V. doi:10.1137/0713009.
  2. 1 2 3 Paige CC, Saunders MA (1981). "Towards a Generalized Singular Value Decomposition". SIAM J. Numer. Anal. 18 (3): 398–405. Bibcode:1981SJNA...18..398P. doi:10.1137/0718026.
  3. Hansen PC (1997). Rank-Deficient and Discrete Ill-Posed Problems: Numerical Aspects of Linear Inversion. SIAM Monographs on Mathematical Modeling and Computation. ISBN   0-89871-403-6.
  4. de Moor BL, Golub GH (1989). "Generalized Singular Value Decompositions A Proposal for a Standard Nomenclauture" (PDF).
  5. de Moor BL, Zha H (1991). "A tree of generalizations of the ordinary singular value decomposition". Linear Algebra and Its Applications. 147: 469–500. doi: 10.1016/0024-3795(91)90243-P .
  6. Alter O, Brown PO, Botstein D (March 2003). "Generalized singular value decomposition for comparative analysis of genome-scale expression data sets of two different organisms". Proceedings of the National Academy of Sciences of the United States of America. 100 (6): 3351–6. Bibcode:2003PNAS..100.3351A. doi: 10.1073/pnas.0530258100 . PMC   152296 . PMID   12631705.
  7. Lee CH, Alpert BO, Sankaranarayanan P, Alter O (January 2012). "GSVD comparison of patient-matched normal and tumor aCGH profiles reveals global copy-number alterations predicting glioblastoma multiforme survival". PLOS ONE. 7 (1): e30098. Bibcode:2012PLoSO...730098L. doi: 10.1371/journal.pone.0030098 . PMC   3264559 . PMID   22291905.
  8. Aiello KA, Ponnapalli SP, Alter O (September 2018). "Mathematically universal and biologically consistent astrocytoma genotype encodes for transformation and predicts survival phenotype". APL Bioengineering. 2 (3): 031909. doi:10.1063/1.5037882. PMC   6215493 . PMID   30397684.
  9. Ponnapalli SP, Bradley MW, Devine K, Bowen J, Coppens SE, Leraas KM, Milash BA, Li F, Luo H, Qiu S, Wu K, Yang H, Wittwer CT, Palmer CA, Jensen RL, Gastier-Foster JM, Hanson HA, Barnholtz-Sloan JS, Alter O (May 2020). "Retrospective Clinical Trial Experimentally Validates Glioblastoma Genome-Wide Pattern of DNA Copy-Number Alterations Predictor of Survival". APL Bioengineering. 4 (2): 026106. doi: 10.1063/1.5142559 . PMC   7229984 . PMID   32478280. Press Release.
  10. Ponnapalli SP, Saunders MA, Van Loan CF, Alter O (December 2011). "A higher-order generalized singular value decomposition for comparison of global mRNA expression from multiple organisms". PLOS ONE. 6 (12): e28072. Bibcode:2011PLoSO...628072P. doi: 10.1371/journal.pone.0028072 . PMC   3245232 . PMID   22216090.
  11. Sankaranarayanan P, Schomay TE, Aiello KA, Alter O (April 2015). "Tensor GSVD of patient- and platform-matched tumor and normal DNA copy-number profiles uncovers chromosome arm-wide patterns of tumor-exclusive platform-consistent alterations encoding for cell transformation and predicting ovarian cancer survival". PLOS ONE. 10 (4): e0121396. Bibcode:2015PLoSO..1021396S. doi: 10.1371/journal.pone.0121396 . PMC   4398562 . PMID   25875127.
  12. Bradley MW, Aiello KA, Ponnapalli SP, Hanson HA, Alter O (September 2019). "GSVD- and tensor GSVD-uncovered patterns of DNA copy-number alterations predict adenocarcinomas survival in general and in response to platinum". APL Bioengineering. 3 (3): 036104. doi:10.1063/1.5099268. PMC   6701977 . PMID   31463421. Supplementary Material.
  13. Cabannes, Vivien; Pillaud-Vivien, Loucas; Bach, Francis; Rudi, Alessandro (2021). "Overcoming the curse of dimensionality with Laplacian regularization in semi-supervised learning". arXiv: 2009.04324 [stat.ML].
  14. Jolliffe IT (2002). Principal Component Analysis . Springer Series in Statistics (2nd ed.). NY: Springer. ISBN   978-0-387-95442-4.
  15. Greenacre M (1983). Theory and Applications of Correspondence Analysis. London: Academic Press. ISBN   978-0-12-299050-2.
  16. Abdi H, Williams LJ (2010). "Principal component analysis". Wiley Interdisciplinary Reviews: Computational Statistics. 2 (4): 433–459. doi:10.1002/wics.101.
  17. Abdi H (2007). "Singular Value Decomposition (SVD) and Generalized Singular Value Decomposition (GSVD).". In Salkind NJ (ed.). Encyclopedia of Measurement and Statistics . Thousand Oaks (CA): Sage. pp.  907–912.

Further reading