Centrala begrepp
A hybrid approach combining Kalman filtering, optimization, and learning-based methods is proposed to accurately estimate the state of a legged robot's trunk by integrating proprioceptive and exteroceptive information.
Sammanfattning
The paper presents a state estimation framework called OptiState that combines a model-based Kalman filter with a learning-based Gated Recurrent Unit (GRU) network to estimate the state of a legged robot's trunk.
The Kalman filter uses joint encoder, IMU, and ground reaction force measurements to provide an initial state estimate. The GRU then refines this estimate by considering the Kalman filter output, latent space representation of depth images from a Vision Transformer (ViT) autoencoder, and other sensor data over a receding time horizon.
The key aspects of the approach are:
- The Kalman filter leverages a single-rigid body model and reuses ground reaction forces from a Model Predictive Control (MPC) optimization to propagate the state.
- The GRU learns to correct any nonlinearities or errors in the Kalman filter's estimate, while also providing an uncertainty measure for the prediction.
- The ViT autoencoder extracts semantic information from depth images to aid the GRU's state estimation.
The proposed OptiState framework is evaluated on a quadruped robot traversing various terrains, including slippery, inclined, and rough surfaces. It demonstrates a 65% improvement in Root Mean Squared Error compared to a state-of-the-art Visual-Inertial Odometry (VIO) SLAM baseline.
Statistik
The robot's trunk state includes orientation (roll, pitch, yaw), position (x, y, z), angular velocity (roll, pitch, yaw), and linear velocity (x, y, z).
The Kalman filter uses joint encoder positions and velocities, IMU orientation and angular velocity, and ground reaction forces from an MPC controller as inputs.
The GRU uses the Kalman filter's state estimate, latent space of depth images, odometry data, IMU accelerations, and ground reaction forces as inputs.
Citat
"By integrating Kalman filtering, optimization, and learning-based modalities, we propose a hybrid solution that combines proprioception and exteroceptive information for estimating the state of the robot's trunk."
"The estimation is further refined through Gated Recurrent Units, which also considers semantic insights and robot height from a Vision Transformer autoencoder applied on depth images."
"This framework not only furnishes accurate robot state estimates, including uncertainty evaluations, but can minimize the nonlinear errors that arise from sensor measurements and model simplifications through learning."