Paired Variational Autoencoders and Transformer-Based Language Models for Robotic Language Learning
A neural model that bidirectionally binds robot actions and their language descriptions using paired variational autoencoders and a pretrained language model (BERT) to enable understanding of unconstrained natural language instructions.