The paper proposes several novel 'order-as-supervision' pre-training methods to improve procedural text understanding, which is challenging due to the changing attributes of entities in the context. The methods include Permutation Classification, Embedding Regression, and Skip-Clip.
The key highlights are:
The proposed methods are evaluated on two downstream Entity Tracking datasets - NPN-Cooking in the recipe domain and ProPara in the open domain. The results show that the order-based pre-training methods outperform baselines and state-of-the-art language models, with improvements of 1.6% and 7-9% across different metrics.
The paper also analyzes the combination of different pre-training strategies, finding that using a single strategy performs better than sequential combinations, as the strategies use different supervision cues.
A otro idioma
del contenido fuente
arxiv.org
Ideas clave extraídas de
by Abhilash Nan... a las arxiv.org 04-09-2024
https://arxiv.org/pdf/2404.04676.pdfConsultas más profundas