The author proposes a method, PMG-AFT, to enhance zero-shot adversarial robustness by leveraging supervision from pre-trained models and clean examples. This approach aims to retain generalization features while mitigating overfitting.
The author explores the sensitivity of text prompts in enhancing adversarial robustness for Vision-Language Models. By proposing Adversarial Prompt Tuning (APT), they demonstrate significant improvements in accuracy and robustness by simply adding one learned word to prompts.