Complex long-horizon and sparse reward robotics tasks are challenging for scaling end-to-end learning methods. By contrast, planning approaches have shown great potential to handle such complex tasks effectively. One of the major criticisms of planning-based approaches has been the lack of availability of accurate world models (aka abstractions) to utilize.
There has been a renewed interest in using learning-based approaches to learn symbolic representations that support planning. However, this research is often fragmented into disjoint sub-communities such as task and motion planning, reinforcement learning (hierarchical, model-based), planning with formal logic, planning with natural language (language models), and neuro-symbolic AI. This workshop aims to create a common forum to share insights, discuss key questions, and chart a path forward via abstraction.
We aim to facilitate this bridge-building in two ways: (1) a diverse selection of papers and invited speakers; and (2) a highly interactive workshop. Concretely, the workshop will highlight approaches that use different learning methods, mainly to learn symbolic and composable representations of world models. Key questions for discussion include:
Following the success of the previous offerings of the workshop at CoRL 2023 and CoRL 2024, we propose a third iteration of the workshop at CoRL 2025. Specifically, the previous iterations of the workshop received a total of 26 and 55 submissions, respectively. These submissions highlighted important characteristics and challenges of learning abstractions while showcasing the abstraction capabilities of pre-trained LLMs and VLMs. This iteration of the workshop would have a stronger emphasis on works that highlight the capabilities of using high-dimensional data and pre-trained models for learning sound and complete abstractions that enable cost-effective and reliable planning.
We solicit papers of the following topics:
We solicit workshop paper submissions relevant to the above call of the following types:
Please format submissions in CoRL or IEEE conference (ICRA or IROS) styles. Submissions do not need to be anonymized. To authors submitting papers rejected from other conferences: please ensure that comments given by the reviewers are addressed prior to submission.
Note: Please feel free to submit work under review or accepted for presentation at other workshops and/or conferences as we will not require copyright transfer.
We are now accepting submissions through our OpenReview portal
Note: The CoRL workshop organizers have requested that we do not accept submissions that are already accepted at the main CoRL conference. We kindly ask authors to respect this policy when submitting to our workshop.
Paper Submission Deadline | Aug 20, 2025 (Early) |
Sep 5, 2025 (Late) | |
Author Notification | Sep 5, 2025 |
Camera-ready Version Due | Sept 22, 2025 |
Workshop | Sept 27, 2025 |
We offer two submission deadlines to accommodate different planning needs: an early deadline and a late deadline. The early deadline is intended for authors who may require quicker notification for travel, visa, or funding arrangements. Submissions received by the early deadline will be reviewed promptly, and notifications will be sent within approximately two weeks of submission. Authors who do not require early feedback are welcome to submit by the late deadline.
Session 1 |
|
1:30 PM - 1:35 PM | Welcome Remarks |
1:35 PM - 2:05 PM | Invited Talk: Danfei Xu |
2:05 PM - 2:35 PM | Invited Talk: Emre Ugur |
2:35 PM - 3:00 PM | Poster Lightning Talks |
3:00 PM - 3:30 PM | Coffee Break + Poster Session |
3:30 PM - 3:55 PM | Poster Session |
3:55 PM - 4:00 PM | Best Paper Award |
4:05 PM - 4:35 PM | Invited Talk: Rohan Paul |
Long-horizon planning is fundamental to our ability to solve complex physical problems, from using tools to cooking dinners. Despite recent progress in commonsense-rich foundation models, the ability to do the same is still lacking in robots. In this talk, I will present a body of work that aims to transform Task and Motion Planning—one of the most powerful computational frameworks in Manipulation Planning—into a fully generative model framework, enabling compositional generalization in a predominantly data-driven approach. I will explore how to chain together modular diffusion-based skills through iterative forward-backward denoising, how to formulate TAMP as a factor graph problem with generative models serving as learned constraints for planning, and how to integrate task and motion planning within a single generative process.
Abstract reasoning are among the most essential characteristics of high-level intelligence that distinguish humans from other animals. If the robots can achieve abstract reasoning on their own, they can perform new tasks in completely novel environments by updating their cognitive skills or by discovering new symbols and rules. Towards this goal, we propose a novel general framework, DeepSym, which discovers interaction grounded, discrete object, action and effect categories and builds probabilistic rules for non-trivial action planning. In DeepSym, our robot interacts with objects using an initial action repertoire and observes the effects it can create in the environment. To form interaction-grounded object, action, effect, and relational categories, we employ a binary bottleneck layer in a predictive, deep encoder-decoder network that takes the image of the scene and the action parameters as input and generates the resulting effects in the scene in pixel coordinates. The knowledge represented by the neural network is distilled into rules and represented in the Probabilistic Planning Domain Definition Language (PPDDL), allowing off-the-shelf planners to operate on the knowledge extracted from the sensorimotor experience of the robot.
Robot instruction following entails translating the human’s intent expressed as high-level language descriptions to contextually ground plans for the robot to execute. The problem is challenging due to the abstract nature of human instructions, large variability of possible tasks and long-horizon spatio-temporal reasoning required for plan synthesis. This talk will discuss recent work in acquiring grounded abstractions for from human annotated demonstrations, leveraging sub-goals in long-horizon skill learning and using VLMs as a source for instruction-specific symbolic knowledge. Overall, we will hope to uncover the role of abstractions in aiding long-range reasoning as well as bridging the human’s intent and robot’s world model.