Complex Interpolation of Matrices with an application to Multi-Manifold Learning
Abstract
Given two symmetric positive-definite matrices $A, B \in \mathbb{R}^{n \times n}$, we study the spectral properties of the interpolation $A^{1-x} B^x$ for $0 \leq x \leq 1$. The presence of common structures' in $A$ and $B$, eigenvectors pointing in a similar direction, can be investigated using this interpolation perspective. Generically, exact log-linearity of the operator norm $\|A^{1-x} B^x\|$ is equivalent to the existence of a shared eigenvector in the original matrices; stability bounds s...
Description / Details
Given two symmetric positive-definite matrices , we study the spectral properties of the interpolation for . The presence of `common structures' in and , eigenvectors pointing in a similar direction, can be investigated using this interpolation perspective. Generically, exact log-linearity of the operator norm is equivalent to the existence of a shared eigenvector in the original matrices; stability bounds show that approximate log-linearity forces principal singular vectors to align with leading eigenvectors of both matrices. These results give rise to and provide theoretical justification for a multi-manifold learning framework that identifies common and distinct latent structures in multiview data.
Source: arXiv:2604.14118v1 - http://arxiv.org/abs/2604.14118v1 PDF: https://arxiv.org/pdf/2604.14118v1 Original Link: http://arxiv.org/abs/2604.14118v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
Apr 17, 2026
Data Science
Machine Learning
0