Leveraging Pretrained Large Language Models for Effective Out-of-Distribution Detection
Pretrained large language models can effectively serve as OOD proxies, and the likelihood ratio between a pretrained LLM and its finetuned variant provides a powerful criterion for detecting out-of-distribution data.