Back to Explorer
Research PaperResearchia:202603.30004[Data Science > Machine Learning]

Tunable Soft Equivariance with Guarantees

Md Ashiqur Rahman

Abstract

Equivariance is a fundamental property in computer vision models, yet strict equivariance is rarely satisfied in real-world data, which can limit a model's performance. Controlling the degree of equivariance is therefore desirable. We propose a general framework for constructing soft equivariant models by projecting the model weights into a designed subspace. The method applies to any pre-trained architecture and provides theoretical bounds on the induced equivariance error. Empirically, we demonstrate the effectiveness of our method on multiple pre-trained backbones, including ViT and ResNet, across image classification, semantic segmentation, and human-trajectory prediction tasks. Notably, our approach improves the performance while simultaneously reducing equivariance error on the competitive ImageNet benchmark.


Source: arXiv:2603.26657v1 - http://arxiv.org/abs/2603.26657v1 PDF: https://arxiv.org/pdf/2603.26657v1 Original Link: http://arxiv.org/abs/2603.26657v1

Submission:3/30/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!