ManipBench: Benchmarking Vision-Language Models for Low-Level Robot Manipulation
Published in Conference on Robot Learning (CoRL), 2025
Propose a novel benchmark, ManipBench, to evaluate the low-level robot manipulation reasoning capabilities of VLMs across various dimensions, including how well they understand object-object interactions and deformable object manipulation Read more
Recommended citation: Enyu Zhao*, Vedant Raval*, Hejia Zhang*, Jiageng Mao, Zeyu Shangguan, Stefanos Nikolaidis, Yue Wang and Daniel Seita. (2025). "ManipBench: Benchmarking Vision-Language Models for Low-Level Robot Manipulation. " Conference on Robot Learning (CoRL).
Download Paper
