The paper proposes several novel 'order-as-supervision' pre-training methods to improve procedural text understanding, which is challenging due to the changing attributes of entities in the context. The methods include Permutation Classification, Embedding Regression, and Skip-Clip.
The key highlights are:
The proposed methods are evaluated on two downstream Entity Tracking datasets - NPN-Cooking in the recipe domain and ProPara in the open domain. The results show that the order-based pre-training methods outperform baselines and state-of-the-art language models, with improvements of 1.6% and 7-9% across different metrics.
The paper also analyzes the combination of different pre-training strategies, finding that using a single strategy performs better than sequential combinations, as the strategies use different supervision cues.
Sang ngôn ngữ khác
từ nội dung nguồn
arxiv.org
Thông tin chi tiết chính được chắt lọc từ
by Abhilash Nan... lúc arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04676.pdfYêu cầu sâu hơn