Core Concepts
An adaptable trustor agent can learn to dynamically toggle between push (CA) and pull (FIRE) trust models to maximize utility in open multi-agent environments with changing conditions.
Abstract
The paper investigates how trustors can detect the presence of dynamic factors in their environment and decide which trust model to employ (CA or FIRE) to maximize utility. This problem is framed as a machine learning problem in a partially observable environment, where the trustor is unaware of the effect of these dynamic factors.
The authors describe how the trustor can calculate various environmental features to assess the current state and then use Deep Q-Learning (DQN) in a single-agent Reinforcement Learning setting to learn the optimal policy - whether to use the push (CA) or pull (FIRE) trust model.
Simulation experiments are conducted to compare the performance of the adaptable trustor (using DQN) with trustors using only one model (FIRE or CA). The results show that the adaptable agent is able to learn when to use each model and demonstrate consistently robust performance in dynamic environments, outperforming the single-model approaches.
The key dynamic factors considered include:
Changes in the provider population
Changes in the consumer population
Providers altering their average performance level
Providers switching to different performance profiles
Consumers moving to new locations
Providers moving to new locations
The trustor estimates the extent of these changes using various features calculated from its local information, such as its rating database, nearby providers, and acquaintances.
Stats
The utility gained (UG) from each interaction is used as the key metric to evaluate the performance of the different trust models.