Core Concepts
A pattern-oriented reference architecture that serves as guidance for designing foundation model-based agents with a focus on ensuring trustworthiness and addressing responsible AI considerations.
Abstract
The paper presents a reference architecture for designing foundation model-based agents. It covers the key components and design patterns required to build such agents, with a strong emphasis on responsible AI principles.
The interaction engineering component focuses on understanding user goals through passive or proactive approaches, and generating appropriate prompts and responses. The memory component manages short-term and long-term information to support the agent's reasoning and decision-making.
The planning component explores single-path and multi-path plan generation strategies, leveraging one-shot or incremental model querying. It also incorporates plan reflection mechanisms for self-improvement.
The execution engine enables task execution, cooperation with other agents or external tools, and task monitoring. Responsible AI plugins are introduced to address key concerns like continuous risk assessment, transparency, and ethical guardrails.
The architecture also discusses the trade-offs in using external foundation models, fine-tuned models, or building sovereign models in-house. The proposed reference architecture is evaluated by mapping it to the architectures of two real-world agents, MetaGPT and HuggingGPT, demonstrating its completeness and utility.