The paper proposes several novel 'order-as-supervision' pre-training methods to improve procedural text understanding, which is challenging due to the changing attributes of entities in the context. The methods include Permutation Classification, Embedding Regression, and Skip-Clip.
The key highlights are:
The proposed methods are evaluated on two downstream Entity Tracking datasets - NPN-Cooking in the recipe domain and ProPara in the open domain. The results show that the order-based pre-training methods outperform baselines and state-of-the-art language models, with improvements of 1.6% and 7-9% across different metrics.
The paper also analyzes the combination of different pre-training strategies, finding that using a single strategy performs better than sequential combinations, as the strategies use different supervision cues.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Abhilash Nan... at arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04676.pdfDeeper Inquiries