Back to Explorer
Research PaperResearchia:202603.11090[Robotics > Robotics]

Bilevel Planning with Learned Symbolic Abstractions from Interaction Data

Fatih Dogangun

Abstract

Intelligent agents must reason over both continuous dynamics and discrete representations to generate effective plans in complex environments. Previous studies have shown that symbolic abstractions can emerge from neural effect predictors trained with a robot's unsupervised exploration. However, these methods rely on deterministic symbolic domains, lack mechanisms to verify the generated symbolic plans, and operate only at the abstract level, often failing to capture the continuous dynamics of the environment. To overcome these limitations, we propose a bilevel neuro-symbolic framework in which learned probabilistic symbolic rules generate candidate plans rapidly at the high level, and learned continuous effect models verify these plans and perform forward search when necessary at the low level. Our experiments on multi-object manipulation tasks demonstrate that the proposed bilevel method outperforms symbolic-only approaches, reliably identifying failing plans through verification, and achieves planning performance statistically comparable to continuous forward search while resolving most problems via efficient symbolic reasoning.


Source: arXiv:2603.08599v1 - http://arxiv.org/abs/2603.08599v1 PDF: https://arxiv.org/pdf/2603.08599v1 Original Link: http://arxiv.org/abs/2603.08599v1

Submission:3/11/2026
Comments:0 comments
Subjects:Robotics; Robotics
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Bilevel Planning with Learned Symbolic Abstractions from Interaction Data | Researchia