The study investigates the relationship between word predictability and reading times across 11 languages from 5 language families. The key findings are:
Surprisal, or negative log probability of a word given its context, is a consistent predictor of reading times across all 11 languages tested. Models that include surprisal as a predictor show significantly better predictive power over baseline models that do not.
Contextual entropy, or the expected surprisal of a word, also contributes to predicting reading times in most languages when added as an additional predictor. However, replacing surprisal with contextual entropy tends to hurt predictive power.
The relationship between surprisal and reading times is found to be linear across languages, contradicting some recent studies that have suggested non-linear relationships.
The magnitude of the surprisal effect, around 2-4 milliseconds per bit of surprisal, is remarkably consistent across languages, suggesting a stable crosslinguistic preference for the rate of information processing during reading.
The predictive power of surprisal shows some variation across languages, but this does not seem to be primarily driven by differences in language model quality across languages.
Overall, the results provide robust crosslinguistic evidence for the role of information-theoretic measures of word predictability in shaping reading behavior, and suggest that the core principles of surprisal theory generalize well beyond English.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Ethan Gotlie... at arxiv.org 09-12-2024
https://arxiv.org/pdf/2307.03667.pdfDeeper Inquiries