Yunzhu Li (@YunzhuLiYZ)

2024-08-29 | โค๏ธ 275 | ๐Ÿ” 48


Super excited to finally release our work, ReKep, a unified task representation using relational keypoint constraints. ๐Ÿค– https://rekep-robot.github.io/

A few key takeaways:

  1. Building on the success of VoxPoser, VLM-generated code has proven to be extremely versatile in task specification.
  2. ReKep provides a much more fine-grained task representation than the value map in VoxPoser.
  3. ReKep nicely connects to the Task and Motion Planning (TAMP) field, characterizing both spatial (e.g., shoe tip alignment) and temporal relationships (e.g., keeping the teapot upright during transport).
  4. ReKep enables real-time motion planning at 10 Hz using an off-the-shelf optimizer, allowing for closed-loop control both within and across different task stages (as shown in the video).
  5. Leveraging modules for their strengths (VLM for task interpretation, constrained optimization for motion planning) remains an effective way to ground foundation models in robotics.
  6. Our investigation complements recent trends in large-scale imitation learning, and I can already see several low-hanging opportunities to combine the best of both worlds!

Check @wenlong_huang โ€˜s thread for a more detailed walkthrough of the work! ๐Ÿ‘

๋ฏธ๋””์–ด

video

Quoted: @wenlong_huang

What structural task representation enables multi-stage, in-the-wild, bimanual, reactive manipulation?

Introducing ReKep: LVM to label keypoints & VLM to write keypoint-based constraints, solve wโ€ฆ


์ธ์šฉ ํŠธ์œ—

Wenlong Huang (@wenlong_huang)

What structural task representation enables multi-stage, in-the-wild, bimanual, reactive manipulation?

Introducing ReKep: LVM to label keypoints & VLM to write keypoint-based constraints, solve w/ optimization for diverse tasks, w/o task-specific training or env models.

๐Ÿงต๐Ÿ‘‡ https://t.co/EnAPYUTjmN

์›๋ณธ ํŠธ์œ—

๐ŸŽฌ ์˜์ƒ

Tags

domain-robotics domain-dev-tools domain-vlm