Back to Explorer
Research PaperResearchia:202603.31023[Data Science > Machine Learning]

Subspace Optimization for Backpropagation-Free Continual Test-Time Adaptation

Damian Sójka

Abstract

We introduce PACE, a backpropagation-free continual test-time adaptation system that directly optimizes the affine parameters of normalization layers. Existing derivative-free approaches struggle to balance runtime efficiency with learning capacity, as they either restrict updates to input prompts or require continuous, resource-intensive adaptation regardless of domain stability. To address these limitations, PACE leverages the Covariance Matrix Adaptation Evolution Strategy with the Fastfood projection to optimize high-dimensional affine parameters within a low-dimensional subspace, leading to superior adaptive performance. Furthermore, we enhance the runtime efficiency by incorporating an adaptation stopping criterion and a domain-specialized vector bank to eliminate redundant computation. Our framework achieves state-of-the-art accuracy across multiple benchmarks under continual distribution shifts, reducing runtime by over 50% compared to existing backpropagation-free methods.


Source: arXiv:2603.28678v1 - http://arxiv.org/abs/2603.28678v1 PDF: https://arxiv.org/pdf/2603.28678v1 Original Link: http://arxiv.org/abs/2603.28678v1

Submission:3/31/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Subspace Optimization for Backpropagation-Free Continual Test-Time Adaptation | Researchia