The paper proposes several novel 'order-as-supervision' pre-training methods to improve procedural text understanding, which is challenging due to the changing attributes of entities in the context. The methods include Permutation Classification, Embedding Regression, and Skip-Clip.
The key highlights are:
The proposed methods are evaluated on two downstream Entity Tracking datasets - NPN-Cooking in the recipe domain and ProPara in the open domain. The results show that the order-based pre-training methods outperform baselines and state-of-the-art language models, with improvements of 1.6% and 7-9% across different metrics.
The paper also analyzes the combination of different pre-training strategies, finding that using a single strategy performs better than sequential combinations, as the strategies use different supervision cues.
לשפה אחרת
מתוכן המקור
arxiv.org
תובנות מפתח מזוקקות מ:
by Abhilash Nan... ב- arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04676.pdfשאלות מעמיקות