Invariant Transformation and Resampling based Epistemic-Uncertainty Reduction
Abstract
An artificial intelligence (AI) model can be viewed as a function that maps inputs to outputs in high-dimensional spaces. Once designed and well trained, the AI model is applied for inference. However, even optimized AI models can produce inference errors due to aleatoric and epistemic uncertainties. Interestingly, we observed that when inferring multiple samples based on invariant transformations of an input, inference errors can show partial independences due to epistemic uncertainty. Leveragi...
Description / Details
An artificial intelligence (AI) model can be viewed as a function that maps inputs to outputs in high-dimensional spaces. Once designed and well trained, the AI model is applied for inference. However, even optimized AI models can produce inference errors due to aleatoric and epistemic uncertainties. Interestingly, we observed that when inferring multiple samples based on invariant transformations of an input, inference errors can show partial independences due to epistemic uncertainty. Leveraging this insight, we propose a "resampling" based inferencing that applies to a trained AI model with multiple transformed versions of an input, and aggregates inference outputs to a more accurate result. This approach has the potential to improve inference accuracy and offers a strategy for balancing model size and performance.
Source: arXiv:2602.23315v1 - http://arxiv.org/abs/2602.23315v1 PDF: https://arxiv.org/pdf/2602.23315v1 Original Link: http://arxiv.org/abs/2602.23315v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
Feb 27, 2026
Artificial Intelligence
AI
0