Core Concepts
Action-bisimulation encoding improves representation learning for reinforcement agents by capturing multi-step controllability.
Abstract
Reinforcement learning agents need to identify relevant state features amidst distractors.
Representation learning filters out irrelevant features, improving sample efficiency.
Action-bisimulation extends single-step controllability with recursive invariance.
Pretraining on action-bisimulation enhances sample efficiency in various environments.
The method captures long-term controllability and control-relevant state features effectively.
Empirical results show the superiority of action-bisimulation over other representation methods.
Background distractors challenge traditional methods but have minimal impact on action-bisimulation.
The encoding captures multi-step relationships and is robust to uncontrollable distractors.
Stats
Myopic controllability captures the moment before a crash but not distant control relevance.
Action-bisimulation extends single-step controllability with recursive invariance constraint.