ExplorerArtificial IntelligenceAI
Research PaperResearchia:202605.04047

Meritocratic Fairness in Budgeted Combinatorial Multi-armed Bandits via Shapley Values

Shradha Sharma

Abstract

We propose a new framework for meritocratic fairness in budgeted combinatorial multi-armed bandits with full-bandit feedback (BCMAB-FBF). Unlike semi-bandit feedback, the contribution of individual arms is not received in full-bandit feedback, making the setting significantly more challenging. To compute arm contributions in BCMAB-FBF, we first extend the Shapley value, a classical solution concept from cooperative game theory, to the $K$-Shapley value, which captures the marginal contribution o...

Submitted: May 4, 2026Subjects: AI; Artificial Intelligence

Description / Details

We propose a new framework for meritocratic fairness in budgeted combinatorial multi-armed bandits with full-bandit feedback (BCMAB-FBF). Unlike semi-bandit feedback, the contribution of individual arms is not received in full-bandit feedback, making the setting significantly more challenging. To compute arm contributions in BCMAB-FBF, we first extend the Shapley value, a classical solution concept from cooperative game theory, to the KK-Shapley value, which captures the marginal contribution of an agent restricted to a set of size at most KK. We show that KK-Shapley value is a unique solution concept that satisfies Symmetry, Linearity, Null player, and efficiency properties. We next propose K-SVFair-FBF, a fairness-aware bandit algorithm that adaptively estimates KK-Shapley value with unknown valuation function. Unlike standard bandit literature on full bandit feedback, K-SVFair-FBF not only learns the valuation function under full feedback setting but also mitigates the noise arising from Monte Carlo approximations. Theoretically, we prove that K-SVFair-FBF achieves O(T3/4)O(T^{3/4}) regret bound on fairness regret. Through experiments on federated learning and social influence maximization datasets, we demonstrate that our approach achieves fairness and performs more effectively than existing baselines.


Source: arXiv:2605.00762v1 - http://arxiv.org/abs/2605.00762v1 PDF: https://arxiv.org/pdf/2605.00762v1 Original Link: http://arxiv.org/abs/2605.00762v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
May 4, 2026
Topic:
Artificial Intelligence
Area:
AI
Comments:
0
Bookmark
Meritocratic Fairness in Budgeted Combinatorial Multi-armed Bandits via Shapley Values | Researchia