Enhancing Domain Generalization through Selective Cross-Modality Distillation with CLIP
Selective Cross-Modality Distillation (SCMD) leverages the capabilities of large vision-language models like CLIP to train a more efficient student model with robust generalization across unseen domains.