Core Concepts
Despite the continuous embedding space being more expressive than the discrete token space, soft prompting and prefix-tuning are potentially less expressive than full fine-tuning, even with the same number of learnable parameters. Prefix-tuning cannot change the relative attention pattern over the content and can only bias the outputs of an attention layer in a fixed direction.
Abstract
The paper analyzes the theoretical capabilities and limitations of context-based fine-tuning techniques, including prompting, in-context learning, soft prompting, and prefix-tuning.
Key highlights:
Soft prompting and prefix-tuning have greater expressiveness than prompting, as they can control the mapping from user input to model output more flexibly.
However, despite this increased expressiveness, prefix-tuning has structural limitations. It cannot change the relative attention over the content tokens and can only bias the output of the attention block in a constant direction, unlike full fine-tuning.
The authors show that while prefix-tuning can effectively elicit skills present in the pretrained model, it may not be able to learn novel tasks that require new attention patterns.
Prefix-tuning can combine skills picked up during pretraining to solve some new tasks similar to pretraining tasks, but it may not learn a completely new task.
The authors also discuss the implications of these findings for model interpretability, catastrophic forgetting, and model alignment.