Back to Explorer
Research PaperResearchia:202603.03051[Data Science > Machine Learning]

Who Guards the Guardians? The Challenges of Evaluating Identifiability of Learned Representations

Shruti Joshi

Abstract

Identifiability in representation learning is commonly evaluated using standard metrics (e.g., MCC, DCI, R^2) on synthetic benchmarks with known ground-truth factors. These metrics are assumed to reflect recovery up to the equivalence class guaranteed by identifiability theory. We show that this assumption holds only under specific structural conditions: each metric implicitly encodes assumptions about both the data-generating process (DGP) and the encoder. When these assumptions are violated, metrics become misspecified and can produce systematic false positives and false negatives. Such failures occur both within classical identifiability regimes and in post-hoc settings where identifiability is most needed. We introduce a taxonomy separating DGP assumptions from encoder geometry, use it to characterise the validity domains of existing metrics, and release an evaluation suite for reproducible stress testing and comparison.


Source: arXiv:2602.24278v1 - http://arxiv.org/abs/2602.24278v1 PDF: https://arxiv.org/pdf/2602.24278v1 Original Link: http://arxiv.org/abs/2602.24278v1

Submission:3/3/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!