This study explores various input representations and prompting strategies for the task of generating market comments from numerical stock price data. The key findings are:
Prompts that resemble programming language expressions, such as Python dictionaries and nested lists, perform better than prompts that are closer to natural language or use longer formats like HTML and LaTeX tables.
The authors hypothesize that prompts with characteristics similar to the text used for pretraining the language models yield better outcomes, as the models can better capture the correspondence between timestamps and prices.
While an existing fine-tuned encoder-decoder model achieves high scores in automatic evaluation metrics, a human evaluation reveals that the prompt-based methods generate comments that are more consistent with the reference.
The study provides insights into effective prompt design for tasks that involve generating text from numerical sequences, which can be extended to other data-to-text generation settings beyond market commentary.
Vers une autre langue
à partir du contenu source
arxiv.org
Questions plus approfondies