Back to Explorer
Research PaperResearchia:202602.24081[Data Science > Machine Learning]

Latent Equivariant Operators for Robust Object Recognition: Promise and Challenges

Minh Dinh

Abstract

Despite the successes of deep learning in computer vision, difficulties persist in recognizing objects that have undergone group-symmetric transformations rarely seen during training-for example objects seen in unusual poses, scales, positions, or combinations thereof. Equivariant neural networks are a solution to the problem of generalizing across symmetric transformations, but require knowledge of transformations a priori. An alternative family of architectures proposes to earn equivariant operators in a latent space from examples of symmetric transformations. Here, using simple datasets of rotated and translated noisy MNIST, we illustrate how such architectures can successfully be harnessed for out-of-distribution classification, thus overcoming the limitations of both traditional and equivariant networks. While conceptually enticing, we discuss challenges ahead on the path of scaling these architectures to more complex datasets.


Source: arXiv:2602.18406v1 - http://arxiv.org/abs/2602.18406v1 PDF: https://arxiv.org/pdf/2602.18406v1 Original Link: http://arxiv.org/abs/2602.18406v1

Submission:2/24/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Latent Equivariant Operators for Robust Object Recognition: Promise and Challenges | Researchia | Researchia