ExplorerData ScienceMachine Learning
Research PaperResearchia:202605.06073

Optimal Posterior Sampling for Policy Identification in Tabular Markov Decision Processes

Cyrille Kone

Abstract

We study the $(\varepsilon, δ)$-PAC policy identification problem in finite-horizon episodic Markov Decision Processes. Existing approaches provide finite-time guarantees for approximate settings ($\varepsilon>0$) but suffer from high computational cost, rendering them hard to implement, and also suffer from suboptimal dependence on $\log(1/δ)$. We propose a randomized and computationally efficient algorithm for best policy identification that combines posterior sampling with an online learning ...

Submitted: May 6, 2026Subjects: Machine Learning; Data Science

Description / Details

We study the (ε,δ)(\varepsilon, δ)-PAC policy identification problem in finite-horizon episodic Markov Decision Processes. Existing approaches provide finite-time guarantees for approximate settings (ε>0\varepsilon>0) but suffer from high computational cost, rendering them hard to implement, and also suffer from suboptimal dependence on log(1/δ)\log(1/δ). We propose a randomized and computationally efficient algorithm for best policy identification that combines posterior sampling with an online learning algorithm to guide exploration in the MDP. Our method achieves asymptotic optimality in sample complexity, also in terms of posterior contraction rate, and runs in O(S2AH)O(S^2AH) per episode, matching standard model-based approaches. Unlike prior algorithms such as MOCA and PEDEL, our guarantees remain meaningful in the asymptotic regime and avoid sub-optimal polynomial dependence on log(1/δ)\log(1/δ). Our results provide both theoretical insights and practical tools for efficient policy identification in tabular MDPs.


Source: arXiv:2605.03921v1 - http://arxiv.org/abs/2605.03921v1 PDF: https://arxiv.org/pdf/2605.03921v1 Original Link: http://arxiv.org/abs/2605.03921v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
May 6, 2026
Topic:
Data Science
Area:
Machine Learning
Comments:
0
Bookmark
Optimal Posterior Sampling for Policy Identification in Tabular Markov Decision Processes | Researchia