A novel Krylov subspace method for approximating Fréchet derivatives of large-scale matrix functions
Abstract
We present a novel Krylov subspace method for approximating , the matrix-vector product of the Fréchet derivative of a large-scale matrix function in direction , a task that arises naturally in the sensitivity analysis of quantities involving matrix functions, such as centrality measures for networks. It also arises in the context of gradient-based methods for optimization problems that feature matrix functions, e.g., when fitting an evolution equation to an observed solution trajectory. In principle, the well-known identity [ f\left( \begin{bmatrix} A & E \ 0 & A \end{bmatrix} \right) \begin{bmatrix} 0 \ \vc{b} \end{bmatrix} = \begin{bmatrix} L_f(A, E) \vc{b} \ f(A) \vc{b} \end{bmatrix}, ] allows one to directly apply any standard Krylov subspace method, such as the Arnoldi algorithm, to address this task. However, this comes with the major disadvantage that the involved block triangular matrix has unfavorable spectral properties, which impede the convergence analysis and, to a certain extent, also the observed convergence. To avoid these difficulties, we propose a novel modification of the Arnoldi algorithm that aims at better preserving the block triangular structure. In turn, this allows one to bound the convergence of the modified method by the best polynomial approximation of the derivative on the numerical range of . Several numerical experiments illustrate our findings.
Source: arXiv:2601.21799v1 - http://arxiv.org/abs/2601.21799v1 PDF: https://arxiv.org/pdf/2601.21799v1 Original Link: http://arxiv.org/abs/2601.21799v1