Details Make a Difference: Object State-Sensitive Neurorobotic Task Planning

Proceedings of the International Conference on Artificial Neural Networks, pages 261--275, doi: 10.1007/978-3-031-72341-4_18 - Sep 2024
Associated documents :  
The state of an object reflects its current status or condition and is important for a robot’s task planning and manipulation. However, detecting an object’s state and generating a state-sensitive plan for robots is challenging. Recently, pre-trained Large Language Models (LLMs) and Vision-Language Models (VLMs) have shown impressive capabilities in generating plans. However, to the best of our knowledge, there is hardly any investigation on whether LLMs or VLMs can also generate object state-sensitive plans. To study this, we introduce an Object State-Sensitive Agent (OSSA), a task-planning agent empowered by pre-trained neural networks. We propose two methods for OSSA: (i) a modular model consisting of a pre-trained vision processing module (dense captioning model, DCM) and a natural language processing model (LLM), and (ii) a monolithic model consisting only of a VLM. To quantitatively evaluate the performances of the two methods, we use tabletop scenarios where the task is to clear the table. We contribute a multimodal benchmark dataset that takes object states into consideration. Our results show that both methods can be used for object state-sensitive tasks, but the monolithic approach outperforms the modular approach. The code for OSSA is available at https://github.com/Xiao-wen-Sun/OSSA

 

@InProceedings{SZLLKW24, 
 	 author =  {Sun, Xiaowen and Zhao, Xufeng and Lee, Jae Hee and Lu, Wenhao and Kerzel, Matthias and Wermter, Stefan},  
 	 title = {Details Make a Difference: Object State-Sensitive Neurorobotic Task Planning}, 
 	 booktitle = {Proceedings of the International Conference on Artificial Neural Networks},
 	 journal = {},
 	 editors = {},
 	 number = {},
 	 volume = {},
 	 pages = {261--275},
 	 year = {2024},
 	 month = {Sep},
 	 publisher = {},
 	 doi = {10.1007/978-3-031-72341-4_18}, 
 }