Name of two different techniques based on the singular value decomposition
In linear algebra, the generalized singular value decomposition (GSVD) is the name of two different techniques based on the singular value decomposition. The two versions differ because one version decomposes two (or more) matrices (much like higher order PCA) and the other version uses a set of constraints imposed on the left and right singular vectors.
The generalized singular value decomposition (GSVD) is a matrix decomposition on a pair of matrices which generalizes the singular value decomposition. It was introduced by Van Loan [1] in 1976 and later developed by Paige and Saunders,[2] which is the version described here. In contrast to the SVD, the GSVD decomposes simultaneously a pair of matrices with the same number of columns. The SVD and the GSVD, as well as some other possible generalizations of the SVD,[3][4][5] are extensively used in the study of the conditioning and regularization of linear systems with respect to quadratic semi-norms. In the following, let , or .
Definition[]
The generalized singular value decomposition of matrices and is
is real diagonal with positive diagonal, and contains the non-zero singular values of in decreasing order,
,
is real non-negative block-diagonal, where with , , and ,
is real non-negative block-diagonal, where with , , and ,
,
,
,
.
We denote , , , and . While is diagonal, is not always diagonal, because of the leading rectangular zero matrix; instead is "bottom-right-diagonal".
Variations[]
There are many variations of the GSVD. These variations are related to the fact that it is always possible to multiply from the left by where is an arbitrary unitary matrix. We denote
, where is upper-triangular and invertible, and is unitary. Such matrices exist by RQ-decomposition.
A generalized singular value of and is a pair such that
We have
By these properties we can show that the generalized singular values are exactly the pairs . We have
Therefore
This expression is zero exactly when and for some .
In,[2] the generalized singular values are claimed to be those which solve . However, this claim only holds when , since otherwise the determinant is zero for every pair ; this can be seen by substituting above.
Generalized inverse[]
Define for any invertible matrix , for any zero matrix , and for any block-diagonal matrix. Then define
It can be shown that as defined here is a generalized inverse of ; in particular a -inverse of . Since it does not in general satisfy , this is not the Moore–Penrose inverse; otherwise we could derive for any choice of matrices, which only holds for certain class of matrices.
Suppose , where and . This generalized inverse has the following properties:
Quotient SVD[]
A generalized singular ratio of and is . By the above properties, . Note that is diagonal, and that, ignoring the leading zeros, contains the singular ratios in decreasing order. If is invertible, then has no leading zeros, and the generalized singular ratios are the singular values, and and are the matrices of singular vectors, of the matrix . In fact, computing the SVD of is one of the motivations for the GSVD, as "forming and finding its SVD can lead to unnecessary and large numerical errors when is ill-conditioned for solution of equations".[2] Hence the sometimes used name "quotient SVD", although this is not the only reason for using GSVD. If is not invertible, then is still the SVD of if we relax the requirement of having the singular values in decreasing order. Alternatively, a decreasing order SVD can be found by moving the leading zeros to the back: , where and are appropriate permutation matrices. Since rank equals the number of non-zero singular values, .
Construction[]
Let
be the SVD of , where is unitary, and and are as described,
, where and ,
, where and ,
by the SVD of , where , and are as described,
by a decomposition similar to a QR-decomposition, where and are as described.
Then
We also have
Therefore
Since has orthonormal columns, . Therefore
We also have for each such that that
Therefore , and
Weighted version[]
The weighted version of the generalized singular value decomposition (GSVD) is a constrained matrix decomposition with constraints imposed on the left and right singular vectors of the singular value decomposition.[6][7][8] This form of the GSVD is an extension of the SVD as such. Given the SVD of an m×n real or complex matrix M
where
Where I is the identity matrix and where and are orthonormal given their constraints ( and ). Additionally, and are positive definite matrices (often diagonal matrices of weights). This form of the GSVD is the core of certain techniques, such as generalized principal component analysis and Correspondence analysis.
The GSVD, formulated as a comparative spectral decomposition,[10] has been successfully applied to signal processing and data science, e.g., in genomic signal processing.[11][12][13]
These applications inspired several additional comparative spectral decompositions, i.e., the higher-order GSVD (HO GSVD)[14] and the tensor GSVD.[15][16]
References[]
^ Van Loan CF (1976). "Generalizing the Singular Value Decomposition". SIAM J. Numer. Anal. 13 (1): 76–83. Bibcode:1976SJNA...13...76V. doi:10.1137/0713009.
^ abcPaige CC, Saunders MA (1981). "Towards a Generalized Singular Value Decomposition". SIAM J. Numer. Anal. 18 (3): 398–405. Bibcode:1981SJNA...18..398P. doi:10.1137/0718026.
^Hansen PC (1997). Rank-Deficient and Discrete Ill-Posed Problems: Numerical Aspects of Linear Inversion. SIAM Monographs on Mathematical Modeling and Computation. ISBN0-89871-403-6.
^Abdi H (2007). "Singular Value Decomposition (SVD) and Generalized Singular Value Decomposition (GSVD).". In Salkind NJ (ed.). Encyclopedia of Measurement and Statistics. Thousand Oaks (CA): Sage. pp. 907–912.