Chen, H., Chen, R., Yi, Y., Quan, X., Li, C., Yan, M., & Zhang, J. (2024). Knowledge Distillation of Black-Box Large Language Models. arXiv preprint arXiv:2401.07013.
This paper addresses the challenge of knowledge distillation from black-box Large Language Models (LLMs) where internal states are inaccessible, aiming to improve the performance of smaller, open-source models by leveraging the capabilities of these powerful, proprietary LLMs.
The researchers propose Proxy-KD, a two-stage method involving: 1) Proxy Model Alignment: A white-box LLM (proxy) is aligned with the black-box teacher LLM through supervised fine-tuning and preference optimization using the Direct Preference Optimization (DPO) algorithm. 2) Student Knowledge Distillation: The student model learns from both hard labels (outputs) of the black-box teacher and soft labels (output distributions) provided by the aligned proxy, incorporating a sample-level weight in the distillation objective to prioritize well-aligned distributions.
Proxy-KD presents a compelling solution for distilling knowledge from advanced black-box LLMs, effectively bridging the gap between proprietary and open-source models. The method leverages the strengths of both black-box and white-box KD while mitigating their limitations.
This research contributes significantly to the field of natural language processing by enabling the development of more capable and efficient open-source LLMs, potentially democratizing access to advanced language processing capabilities.
The study acknowledges limitations regarding training time overhead due to proxy alignment and preference optimization. Future research could explore more efficient alignment strategies and investigate the impact of different proxy model architectures and sizes on distillation effectiveness. Additionally, expanding experiments to include other LLM backbones beyond the Llama series would provide a more comprehensive evaluation of Proxy-KD's generalizability.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Hongzhan Che... at arxiv.org 11-12-2024
https://arxiv.org/pdf/2401.07013.pdfDeeper Inquiries