Back to Explorer
Research PaperResearchia:202602.20065[Artificial Intelligence > AI]

Almost Sure Convergence of Differential Temporal Difference Learning for Average Reward Markov Decision Processes

Ethan Blaser

Abstract

The average reward is a fundamental performance metric in reinforcement learning (RL) focusing on the long-run performance of an agent. Differential temporal difference (TD) learning algorithms are a major advance for average reward RL as they provide an efficient online method to learn the value functions associated with the average reward in both on-policy and off-policy settings. However, existing convergence guarantees require a local clock in learning rates tied to state visit counts, which practitioners do not use and does not extend beyond tabular settings. We address this limitation by proving the almost sure convergence of on-policy nn-step differential TD for any nn using standard diminishing learning rates without a local clock. We then derive three sufficient conditions under which off-policy nn-step differential TD also converges without a local clock. These results strengthen the theoretical foundations of differential TD and bring its convergence analysis closer to practical implementations.


Source: arXiv:2602.16629v1 - http://arxiv.org/abs/2602.16629v1 PDF: https://arxiv.org/pdf/2602.16629v1 Original Link: http://arxiv.org/abs/2602.16629v1

Submission:2/20/2026
Comments:0 comments
Subjects:AI; Artificial Intelligence
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!