TY - GEN
T1 - Pseudospectral shattering, the sign function, and diagonalization in nearly matrix multiplication time
AU - Banks, Jess
AU - Garza-Vargas, Jorge
AU - Kulkarni, Archit
AU - Srivastava, Nikhil
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/11
Y1 - 2020/11
N2 - We exhibit a randomized algorithm which given a square matrix A in mathbb{C}{n times n} with Vert A Vert leq 1 and delta > 0, computes with high probability an invertible V and diagonal D such that begin{equation} Vert A-VDV{-1} Vert leq delta end{equation} in O(T {text{MM}}(n) log{2}(n delta)) arithmetic operations on a floating point machine with O(log{4}(n delta) log n) bits of precision. The computed similarity V additionally satisfies Vert V Vert Vert V{-1} Vert leq O(n{2.5} delta). Here T {text{MM}}(n) is the number of arithmetic operations required to multiply two n times n complex matrices numerically stably, known to satisfy T {text{MM}}(n)=O(n{omega+ eta}) for every eta > 0 where omega is the exponent of matrix multiplication [1]. The algorithm is a variant of the spectral bisection algorithm in numerical linear algebra [2] with a crucial Gaussian perturbation preprocessing step. Our running time is optimal up to polylogarithmic factors, in the sense that verifying that a given similarity diagonalizes a matrix requires at least matrix multiplication time. It significantly improves the previously best known provable running times of O(n{10} delta{2}) arithmetic operations for diagonalization of general matrices [3], and (with regards to the dependence on n) O(n{3}) arithmetic operations for Hermitian matrices [4], and is the first algorithm to achieve nearly matrix multiplication time for diagonalization in any model of computation (real arithmetic, rational arithmetic, or finite arithmetic). The proof rests on two new ingredients. (1) We show that adding a small complex Gaussian perturbation to any matrix splits its pseudospectrum into n small well-separated components. In particular, this implies that the eigenvalues of the perturbed matrix have a large minimum gap, a property of independent interest in random matrix theory. (2) We give a rigorous analysis of Roberts' [5] Newton iteration method for computing the sign function of a matrix in finite arithmetic, itself an open problem in numerical analysis since at least 1986 [6]. This is achieved by controlling the evolution of the pseudospectra of the iterates using a carefully chosen sequence of shrinking contour integrals in the complex plane.
AB - We exhibit a randomized algorithm which given a square matrix A in mathbb{C}{n times n} with Vert A Vert leq 1 and delta > 0, computes with high probability an invertible V and diagonal D such that begin{equation} Vert A-VDV{-1} Vert leq delta end{equation} in O(T {text{MM}}(n) log{2}(n delta)) arithmetic operations on a floating point machine with O(log{4}(n delta) log n) bits of precision. The computed similarity V additionally satisfies Vert V Vert Vert V{-1} Vert leq O(n{2.5} delta). Here T {text{MM}}(n) is the number of arithmetic operations required to multiply two n times n complex matrices numerically stably, known to satisfy T {text{MM}}(n)=O(n{omega+ eta}) for every eta > 0 where omega is the exponent of matrix multiplication [1]. The algorithm is a variant of the spectral bisection algorithm in numerical linear algebra [2] with a crucial Gaussian perturbation preprocessing step. Our running time is optimal up to polylogarithmic factors, in the sense that verifying that a given similarity diagonalizes a matrix requires at least matrix multiplication time. It significantly improves the previously best known provable running times of O(n{10} delta{2}) arithmetic operations for diagonalization of general matrices [3], and (with regards to the dependence on n) O(n{3}) arithmetic operations for Hermitian matrices [4], and is the first algorithm to achieve nearly matrix multiplication time for diagonalization in any model of computation (real arithmetic, rational arithmetic, or finite arithmetic). The proof rests on two new ingredients. (1) We show that adding a small complex Gaussian perturbation to any matrix splits its pseudospectrum into n small well-separated components. In particular, this implies that the eigenvalues of the perturbed matrix have a large minimum gap, a property of independent interest in random matrix theory. (2) We give a rigorous analysis of Roberts' [5] Newton iteration method for computing the sign function of a matrix in finite arithmetic, itself an open problem in numerical analysis since at least 1986 [6]. This is achieved by controlling the evolution of the pseudospectra of the iterates using a carefully chosen sequence of shrinking contour integrals in the complex plane.
KW - Numerical Analysis
KW - Random Matrix Theory
UR - https://www.scopus.com/pages/publications/85100332006
UR - https://www.scopus.com/pages/publications/85100332006#tab=citedBy
U2 - 10.1109/FOCS46700.2020.00056
DO - 10.1109/FOCS46700.2020.00056
M3 - Conference contribution
AN - SCOPUS:85100332006
T3 - Proceedings - Annual IEEE Symposium on Foundations of Computer Science, FOCS
SP - 529
EP - 540
BT - Proceedings - 2020 IEEE 61st Annual Symposium on Foundations of Computer Science, FOCS 2020
PB - IEEE Computer Society
T2 - 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020
Y2 - 16 November 2020 through 19 November 2020
ER -