A Randomized Numerical Method for Joint Eigenvalues of Commuting Matrices
Haoze He, Daniel Kressner, Bor Plestenjak
Abstract
Let \({\cal A}=\{A_1,\ldots ,A_d\}\) be a commuting family of \(n\times n\) complex matrices, i.e., \(A_jA_k=A_kA_j\) for all \(1\le j,k\le d\). Then there exists a unitary matrix \(U\) such that all matrices \(U^*A_1U,\ldots ,U^*A_dU\) are upper triangular and the \(n\) \(d\)-tuples containing the diagonal elements of \(U^*A_1U,\ldots ,U^*A_dU\) are called the joint eigenvalues of \(\cal A\). For every joint eigenvalue \(\blambda =(\lambda _1,\ldots ,\lambda _d)\) of \(\cal A\) there exists a nonzero common eigenvector \(x\), such that \(A_ix=\lambda _i x\) for \(i=1,\ldots ,d\).
The task of numerical computation of joint eigenvalues for a commuting family arises, e.g., in solvers for multiparameter eigenvalue problems and systems of multivariate polynomials. We propose and analyze a simple approach, summarized in Algorithm 1, that computes eigenvalues as one-sided or two-sided Rayleigh quotients from eigenvectors of a random linear combination
\(\seteqnumber{0}{}{0}\)\begin{equation} \label {eq:Amu} A(\mu )=\mu _1 A_1 + \mu _2 A_2 + \cdots + \mu _d A_d, \end{equation}
where \(\mu =[\mu _1\ \cdots \ \mu _d]^T\) is a random vector from the uniform distribution on the unit sphere in \(\mathbb C^d\).
We show that Algorithm 1, in particular the use of two-sided Rayleigh quotients, accurately computes well-conditioned semisimple joint eigenvalues with high probability. It still works satisfactorily in the presence of defective eigenvalues. Experiments show that the method can be efficiently used in solvers for multiparameter eigenvalue problems and roots of systems of multivariate polynomials.
Input: A nearly commuting family \({\mathcal {A}}= \{A_1,\ldots , A_d\}\), \(\text {opt} \in \{\text {RQ1},\text {RQ2}\}\).
Output: Approximations of joint eigenvalues of \(\mathcal {A}\).
-
1: Draw \(\mu \in \mathbb {C}^d\) from the uniform distribution on the unit sphere.
-
2: Compute \(A(\mu ) = \mu _1 A_1 + \cdots + \mu _d A_d\).
-
3: Compute invertible matrices \(X, Y\) such that the columns of \(X\) have norm \(1\), \(Y^* X = I\),
-
4: and \(Y^* A(\mu ) X\) is diagonal.
-
5: if opt \(=\) RQ\(1\) then return \({\blambda }_{ \mathrm {RQ1}}^{(i)} = (x_i^* A_1 x_i, \ldots , x_i^* A_d x_i), \quad i = 1,\ldots ,n\).
-
0.2em
-
6: else if opt \(=\) RQ\(2\) then return \({\blambda }_{ \mathrm {RQ2}}^{(i)} = (y_i^* A_1 x_i, \ldots , y_i^* A_d x_i), \quad i = 1,\ldots ,n\).
-
7: end if
The idea of using a random linear combination like (1) is not new. For example, in [1, 4] the unitary matrix \(U\) from the Schur decomposition \(A(\mu )=U^*RU\) is used to transform all matrices from \(\cal A\) to block upper triangular form. Using the Schur decomposition, however, requires clustering to group multiple eigenvalues together, and this is a numerically subtle task. On the other hand, Algorithm 1 does not require clustering and in practice often leads to equally good or even better results for, e.g., multiparameter eigenvalue problems [5] and multivariate root finding problems.
For a significantly simpler situation of commuting Hermitian matrices, where a unitary matrix exists that jointly diagonalizes all matrices, randomized methods based on (1) have recently been analyzed in [2], establishing favorable robustness and stability properties.
An important source of joint eigenvalue problems are eigenvector-based methods for solving systems of multivariate polynomial equations. If we are looking for roots of a set of polynomials
\(\seteqnumber{0}{}{1}\)\begin{equation} \label {eq:polynomials} p_i(x_1,\ldots ,x_d)=0,\quad i=1,\ldots ,m, \end{equation}
such that the solution consists of finitely many points, then a common feature of these methods is that they construct so called multiplication matrices \(M_{x_1},\ldots ,M_{x_d}\) that commute and their joint eigenvalues are the roots \((x_1,\ldots ,x_d)\) of (2). Many techniques that use symbolic and/or numerical computation, including Gröbner basis, various resultants, and Macaulay matrices, are used to construct the multiplication matrices, see, e.g., [6].
Another source are multiparameter eigenvalue problems. A \(d\)-parameter version has the form
\(\seteqnumber{0}{}{2}\)\begin{equation} \label {eq:mep} A_{i0}x_i = \lambda _1 A_{i1}x_i + \cdots + \lambda _d A_{id}x_i, \quad i=1,\ldots ,d, \end{equation}
where \(A_{ij}\) is an \(n_i\times n_i\) complex matrix and \(x_i\ne 0\) for \(i=1,\ldots ,d\). When (3) is satisfied, a \(d\)-tuple \(\blambda =(\lambda _1,\ldots ,\lambda _d)\in \CC ^d\) is called an eigenvalue and \(x_1\otimes \cdots \otimes x_d\) is a corresponding eigenvector. Generically, a multiparameter eigenvalue problem (3) has \(N=n_1\cdots n_d\) eigenvalues. The problem (3) is related to a system of \(d\) generalized eigenvalue problems
\[ \Delta _i z =\lambda _i \Delta _0z, \quad i=1,\ldots ,d, \]
with \(z=x_1\otimes \cdots \otimes x_d\) and the \(N\times N\) matrices (that are called operator determinants)
\[\Delta _0=\left |\begin {matrix}A_{11} & \cdots & A_{1d}\cr \vdots & & \vdots \cr A_{d1} & \cdots & A_{dd}\end {matrix}\right |_\otimes ,\quad \Delta _i=\left |\begin {matrix}A_{11} & \cdots & A_{1,i-1} & A_{10} & A_{1,i+1} & \cdots & A_{1d}\cr \vdots & & \vdots & \vdots & \vdots & & \vdots \cr A_{d1} & \cdots & A_{d,i-1} & A_{d0} & A_{d,i+1} & \cdots & A_{dd}\end {matrix}\right |_\otimes ,\quad i=1,\ldots ,d. \]
If \(\Delta _0\) is invertible, then the matrices \(\Gamma _i:=\Delta _0^{-1}\Delta _i\) for \(i=1,\ldots ,d\) commute. If \(N\) is not too large, then a standard approach to solve (3) is to explicitly compute the matrices \(\Gamma _1,\ldots ,\Gamma _d\) and then solve the joint eigenvalue problem.
References
-
[1] R.M. Corless, P.M Gianni, B.M. Trager, A reordered Schur factorization method for zero dimensional polynomial systems with multiple roots, in Proceedings of the 1997 International Symposium on Symbolic and Algebraic Computation (Kihei, HI) 133–140, ACM, New York, 1997.
-
[2] H. He, D. Kressner, Randomized joint diagonalization of symmetric matrices, SIAM J. Matrix Anal. Appl. 45 (2024) 661–684.
-
[3] H. He, D. Kressner, B. Plestenjak, Randomized methods for computing joint eigenvalues, with applications to multiparameter eigenvalue problems and root finding, arXiv:2409.00500 (2024), to appear in Numer. Algorithms.
-
[4] D. Manocha, J. Demmel, Algorithms for intersecting parametric and algebraic curves I: simple intersections, ACM Trans. Graph. 13 (1994) 73–100.
-
[5] B. Plestenjak, MultiParEig. Toolbox for multiparameter and singular eigenvalue problems, www.mathworks.com/matlabcentral/fileexchange/47844-multipareig, MATLAB Central File Exchange.
-
[6] H. J. Stetter, Numerical Polynomial Algebra, SIAM, Philadelphia, PA, 2004.