Core Concepts
The Fermi-Bose Machine (FBM) proposes a local contrastive learning approach to learn semantically meaningful and adversarially robust representations by geometrically separating the internal representations of neural networks.
Abstract
The Fermi-Bose Machine (FBM) is a novel representation learning framework that aims to address two key challenges in deep neural networks: the lack of biological plausibility in backpropagation training, and the adversarial vulnerability of standard neural networks.
The key idea of FBM is to design a local contrastive learning approach inspired by the concepts of fermions and bosons in physics. In this framework, the representations for inputs belonging to the same class are encouraged to "shrink" (like bosons), while the representations for inputs from different classes are forced to "repel" (like fermions). This layer-wise learning is local in nature and biologically plausible.
The authors provide a statistical mechanics analysis of the FBM, which reveals that the target fermion-pair distance is a key parameter that controls the geometry of the learned representations. Specifically, the analysis shows that by tuning this target distance, the FBM can achieve a "double ascent" phenomenon in generalization accuracy, and more importantly, can significantly improve the adversarial robustness of the neural network without the need for adversarial training.
When applied to the MNIST dataset, the FBM demonstrates intriguing geometric properties of the learned representations, such as the emergence of compact and well-separated prototype manifolds. This geometric structure, facilitated by the local contrastive learning, enables the FBM to outperform the standard multilayer perceptron (MLP) trained with backpropagation, both in terms of generalization accuracy and adversarial robustness.
The authors argue that the FBM provides a principled route towards aligning machine learning with human cognition processing, by learning semantically meaningful and disentangled representations that are also robust to adversarial perturbations.
Stats
The dataset used in the experiments is the MNIST handwritten digit dataset, which consists of 60,000 training images and 10,000 test images.
Quotes
"Distinct from human cognitive processing, deep neural networks trained by backpropagation can be easily fooled by adversarial examples."
"We resolve these two challenges by a unified principle formulated as a statistical mechanics problem."
"Remarkably, without adversarial training, our FBM demonstrates the ability to mitigate the adversarial vulnerability by tuning only a target fermion-pair distance."