Provisioning Large Language Model Agents for Edge Intelligence in SAGINs
The author proposes a joint caching and inference framework for sustainable and ubiquitous LLM agents in SAGINs, introducing the concept of "cached model-as-a-resource" to optimize provisioning. The approach involves optimizing model caching and inference to enhance allocation efficiency while ensuring strategy-proofness.