Kernekoncepter
Platypus family of fine-tuned LLMs excels in performance and efficiency.
Statistik
13B Platypus model can be trained on a single A100 GPU using 25k questions in 5 hours.
Citater
"Scaling laws for neural language models." - Kaplan et al., 2020
"Outrageously large neural networks: The sparsely-gated mixture-of-experts layer." - Shazeer et al., 2017
"LoRA: Low-rank adaptation of large language models." - Hu et al., 2022