Core Concepts
Language models can achieve human-like abstraction of grammatical gender through few-shot learning of novel nouns.
Abstract
Humans and language models exhibit biases in gender categorization. Both can generalize grammatical gender to new words with few examples, but show a bias towards masculine gender. Language models update embeddings for novel nouns, suggesting an abstract representation of gender. Human participants also display a masculine bias and struggle with one-shot learning of novel noun genders. Further research is needed to understand the mechanisms underlying these biases and learning patterns.
Stats
Language models can predict gender agreement with accuracies above chance (50%) across different constructions.
Both LSTM and transformer models exhibit biases towards masculine gender in baseline tasks and few-shot learning experiments.
Weight changes during few-shot learning primarily affect the embeddings of the novel noun and related words from the learning examples.