This research paper introduces a novel prompt tuning method for Vision-Language Models (VLMs) that enhances their ability to generalize to unseen classes while maintaining strong performance on seen classes.
Increasing the information density of prompts can significantly improve the generalization ability of vision-language models, while drastically reducing the number of tunable parameters.