Core Concepts
Using large language models for translating natural language task descriptions to formal task specifications improves task success rates in complex environments.
Abstract
The content discusses the use of Large Language Models (LLMs) for translating natural language instructions into formal task specifications for robots. The approach involves autoregressive translation from high-level task descriptions to an intermediate representation, Signal Temporal Logic (STL), which is then used by a Task-and-Motion Planning (TAMP) algorithm. By addressing syntactic and semantic errors through re-prompting techniques, the method outperforms direct LLM planning in challenging 2D task domains. A comprehensive experimental evaluation across various scenarios demonstrates the effectiveness of the proposed AutoTAMP framework.
I. Introduction
Effective human-robot interaction requires understanding, planning, and executing complex tasks.
Recent advances in Large Language Models (LLMs) show promise for translating natural language into robot action sequences.
Existing approaches face challenges with complex environmental and temporal constraints.
II. Problem Description
Conversion of natural language instructions into motion plans encoded as timed waypoints.
Environment state described by named obstacles provided as additional context.
Generation of constraint-satisfying trajectories based on instructions and environment state.
III. Methods
Comparison of different approaches using LLMs for task planning.
Introduction of AutoTAMP framework for translating NL to STL and planning trajectories.
Utilization of re-prompting techniques for correcting syntax and semantic errors.
IV. Experimental Design
Evaluation across single-agent and multi-agent scenarios with varying constraints.
Impact assessment of error correction on translation performance.
Integration of NL2TL model for comparison with pre-trained LLMs.
V. Results
Task success rates compared across different methods using GPT-3 and GPT-4 as LLMs.
Failures analyzed based on execution time violations, action sequencing issues, and translation errors.
VI. Related Work
Overview of related research on combined task and motion planning, LLMs in TAMP, and NL-to-task representation mapping.
VII. Conclusion
AutoTAMP framework shows improved performance over direct LLM planning in handling complex geometric and temporal constraints.
Stats
"We show that our approach outperforms several methods using LLMs as planners in complex task domains."
"We conduct an ablation study over the translation step by integrating a fine-tuned NL-to-STL model."
"The cost of planning time is high, especially when there are multiple iterations of re-prompting."
Quotes
"We conclude that in-context learning with pre-trained LLMs is well suited for language-to-task-specification translation."
"Our work addresses some limitations of prior approaches."
"Translation with no error correction has modest success across task scenarios."