Skip to content

Commit 90ed54a

Browse files
committed
agenticlab
1 parent 63ca718 commit 90ed54a

3 files changed

Lines changed: 191 additions & 140 deletions

File tree

_bibliography/papers.bib

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,7 @@ @string{iros
66
@string{ral = {IEEE Robotics and Automation Letters}}
77
@string{tro = {IEEE Transactions on Robotics}}
88
@string{ijrr = {The International Journal of Robotics Research}}
9+
910
@inproceedings{huangjadhav2025prrtc,
1011
title = {{pRRTC}: {GPU}-Parallel {RRT}-Connect for Fast, Consistent, and Low-Cost Motion Planning},
1112
author = {Chih H. Huang* and Pranav Jadhav* and Brian Plancher and Zachary Kingston},
@@ -89,6 +90,20 @@ @article{guo2026stac
8990
abbr = {RAL},
9091
preview = {stac.jpg}
9192
}
93+
@misc{guo2026agenticlab,
94+
title = {{AgenticLab}: A Real-World Robot Agent Platform that Can See, Think, and Act},
95+
author = {Pengyuan Guo and Zhonghao Mai and Zhengtong Xu and Kaidi Zhang and Heng Zhang and Zichen Miao and Arash Ajoudani and Zachary Kingston and Qiang Qiu and Yu She},
96+
abstract = {Recent advances in large vision-language models (VLMs) have demonstrated generalizable open-vocabulary perception and reasoning, yet their real-robot manipulation capability remains unclear for long-horizon, closed-loop execution in unstructured, in-the-wild environments. Prior VLM-based manipulation pipelines are difficult to compare across different research groups' setups, and many evaluations rely on simulation, privileged state, or specially designed setups. We present AgenticLab, a model-agnostic robot agent platform and benchmark for open-world manipulation. AgenticLab provides a closed-loop agent pipeline for perception, task decomposition, online verification, and replanning. Using AgenticLab, we benchmark state-of-the-art VLM-based agents on real-robot tasks in unstructured environments. Our benchmark reveals several failure modes that offline vision-language tests (e.g., VQA and static image understanding) fail to capture, including breakdowns in multi-step grounding consistency, object grounding under occlusion and scene changes, and insufficient spatial reasoning for reliable manipulation. We will release the full hardware and software stack to support reproducible evaluation and accelerate research on general-purpose robot agents.},
97+
year = 2026,
98+
eprint = {2602.01662},
99+
archivePrefix = {arXiv},
100+
primaryClass = {cs.RO},
101+
pdf = {https://arxiv.org/abs/2602.01662},
102+
note = {Under Review},
103+
projects = {implicit,long-horizon},
104+
abbr = {ARXIV},
105+
preview = {agenticlab.gif}
106+
}
92107
@misc{yan2025vizcoast,
93108
title = {Using {VLM} Reasoning to Constrain Task and Motion Planning},
94109
author = {Muyang Yan* and Miras Mengdibayev* and Ardon Floros and Weihang Guo and Lydia E. Kavraki and Zachary Kingston},

0 commit comments

Comments
 (0)