AutoCGP: Closed-Loop Concept-Guided Policies from Unlabeled Demonstrations

2citations
2
Citations
#1476
in ICLR 2025
of 3827 papers
6
Authors
1
Data Points

Abstract

Training embodied agents to perform complex robotic tasks presents significant challenges due to the entangled factors of task compositionality, environmental diversity, and dynamic changes. In this work, we introduce a novel imitation learning framework to train closed-loop concept-guided policies that enhance long-horizon task performance by leveraging discovered manipulation concepts. Unlike methods that rely on predefined skills and human-annotated labels, our approach allows agents to autonomously abstract manipulation concepts from their proprioceptive states, thereby alleviating misalignment due to ambiguities in human semantics and environmental complexity. Our framework comprises two primary components: anAutomatic Concept Discoverymodule that identifies meaningful and consistent manipulation concepts, and aConcept-Guided Policy Learningmodule that effectively utilizes these manipulation concepts for adaptive task execution, including aConcept Selection Transformerfor concept-based guidance and aConcept-Guided Policyfor action prediction with the selected concepts. Experiments demonstrate that our approach significantly outperforms baseline methods across a range of tasks and environments, while showcasing emergent consistency in motion patterns associated with the discovered manipulation concepts. Codes are available at: https://github.com/PeiZhou26/AutoCGP.

Citation History

Jan 25, 2026
2