Concetti Chiave
This paper presents a novel method for detecting landmarks to decompose complex tasks into subtasks, leveraging first-order logic state representation and large language models (LLMs) to create interpretable rule-based policies through an inductive logic programming (ILP)-based reinforcement learning agent.
Sintesi
The paper introduces a novel approach for identifying landmarks and decomposing complex tasks into subtasks, which are necessary states or subsets of states that must be visited to complete a task. The authors use a contrastive learning algorithm to detect potential landmark states, followed by a graph search algorithm to identify the necessary grounded predicates for each subtask.
To generate rules for achieving the identified subtasks, the authors leverage the common sense knowledge embedded within LLMs. They use few-shot learning with the LLAMA 3.1 model to generate rule templates, which are then further refined and used as input for an ILP-based RL agent to formulate a rule-based policy.
The experiments demonstrate the accuracy and efficiency of the landmark identification algorithm, as well as the effectiveness of the LLM-guided rule generation in solving complex tasks. The authors show that their approach reduces reliance on predefined logic predicates, offering a more flexible and scalable solution compared to previous methods that rely on human-generated rule templates.
The key highlights of the paper include:
- A novel algorithm for identifying landmarks and decomposing complex tasks into subtasks using contrastive learning and graph search.
- Leveraging LLMs to efficiently generate rule templates for achieving the identified subtasks, reducing the need for predefined logic predicates.
- Integrating the LLM-generated rules with an ILP-based RL agent to formulate an interpretable rule-based policy.
- Experimental results showcasing the accuracy, efficiency, and effectiveness of the proposed approach in solving complex tasks.
Statistiche
One of the fundamental challenges in reinforcement learning is to take a complex task and be able to decompose it to subtasks that are simpler for the RL agent to learn.
Our experiments show that our LLM-guided rule template generation can produce rules that are necessary for solving a subtask, which leads to solving complex tasks with fewer assumptions about predefined first-order logic predicates of the environment.
The results of the algorithm are detailed in the experimental section of this paper.
Our graph search algorithm identifies the largest set of predicates that reliably activate landmarks, treated as subtasks for the next stage.
The experimental results demonstrate that the algorithm is both accurate and efficient in subtask detection and that LLM-guided rule generation reduces reliance on predefined logic predicates, offering a more flexible and scalable solution.
Citazioni
"One of the fundamental challenges in reinforcement learning RL is to take a complex task and be able to decompose it to subtasks that are simpler for the RL agent to learn."
"Our experiments show that our LLM-guided rule template generation can produce rules that are necessary for solving a subtask, which leads to solving complex tasks with fewer assumptions about predefined first-order logic predicates of the environment."
"The results of the algorithm are detailed in the experimental section of this paper."
"Our graph search algorithm identifies the largest set of predicates that reliably activate landmarks, treated as subtasks for the next stage."
"The experimental results demonstrate that the algorithm is both accurate and efficient in subtask detection and that LLM-guided rule generation reduces reliance on predefined logic predicates, offering a more flexible and scalable solution."