WildLMa: Long Horizon Loco-Manipulation in the Wild
Abstract
`In-the-wild' mobile manipulation aims to deploy robots in diverse real-world environments, which requires the robot to (1) have skills that generalize across object configurations; (2) be capable of long-horizon task execution in diverse environments; and (3) perform complex manipulation beyond pick-and-place. Quadruped robots with manipulators hold promise for extending the workspace and enabling robust locomotion, but existing results do not investigate such a capability. This paper proposes WildLMa with three components to address these issues: (1) adaptation of learned low-level controller for VR-enabled whole-body teleoperation and traversability; (2) WildLMa-Skill -- a library of generalizable visuomotor skills acquired via imitation learning or heuristics and (3) WildLMa-Planner -- an interface of learned skills that allow LLM planners to coordinate skills for long-horizon tasks. We demonstrate the importance of high-quality training data by achieving higher grasping success rate over existing RL baselines using only tens of demonstrations. WildLMa exploits CLIP for language-conditioned imitation learning that empirically generalizes to objects unseen in training demonstrations. Besides extensive quantitative evaluation, we qualitatively demonstrate practical robot applications, such as cleaning up trash in university hallways or outdoor terrains, operating articulated objects, and rearranging items on a bookshelf.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Generalizable Humanoid Manipulation with Improved 3D Diffusion Policies (2024)
- Helpful DoggyBot: Open-World Object Fetching using Legged Robots and Vision-Language Models (2024)
- QuadWBG: Generalizable Quadrupedal Whole-Body Grasping (2024)
- Precise and Dexterous Robotic Manipulation via Human-in-the-Loop Reinforcement Learning (2024)
- LEGATO: Cross-Embodiment Imitation Using a Grasping Tool (2024)
- Single-Shot Learning of Stable Dynamical Systems for Long-Horizon Manipulation Tasks (2024)
- DexH2R: Task-oriented Dexterous Manipulation from Human to Robots (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper