Neuroscience-inspired motion energy models outperform state-of-the-art deep learning optical flow models in zero-shot generalization to random dot stimuli for object segmentation, suggesting a closer alignment with human visual processing.
This paper introduces a novel unsupervised method for segmenting moving objects from aerial platforms using event cameras and self-supervised vision transformers, achieving state-of-the-art performance on multiple benchmarks.