Core Concepts
Computational metrics like sentence surprisal and relevance predict human sentence comprehension effectively across languages.
Abstract
Research focuses on computational metrics for predicting human sentence comprehension.
Introduces methods like surprisal and semantic relevance to understand sentence processing.
Utilizes multilingual large language models (LLMs) for accurate predictions.
Attention-aware approach enhances contextual information computation.
Results show the effectiveness of these metrics in predicting reading speed across languages.
Stats
Word surprisal estimates information among words in context (Hale, 2001).
Surprisal computed by neural language models underpredicts human reading times (Van Schijndel and Linzen, 2021).
Semantic similarity gauges the similarity between meanings of words or phrases (Mitchell and Lapata, 2010).
Quotes
"Prediction could be a key mechanism underlying human language comprehension."
"Relevant sentences tend to be processed more quickly, creating discourse coherence."
"Combining surprisal and relevance can better simulate human language processing."