Khái niệm cốt lõi
This work introduces the problem of Causally Abstracted Multi-armed Bandits (CAMABs), where decision-making problems are modeled at different levels of resolution and related via causal abstraction. The authors propose algorithms to learn across these related models and analyze their regret.
Tóm tắt
The paper introduces the problem of Causally Abstracted Multi-armed Bandits (CAMABs), where decision-making problems are modeled at different levels of resolution and related via causal abstraction. The key insights are:
Rewriting the title to be more informative and honest.
The content is categorized as Causal Inference.
The main topic is Causally Abstracted Multi-armed Bandits.
The core message is that the authors introduce the CAMAB problem, where decision-making problems are modeled at different levels of resolution and related via causal abstraction, and propose algorithms to learn across these related models.
The paper first provides background on Structural Causal Models (SCMs) and Causal Multi-armed Bandits (CMABs). It then defines the CAMAB problem, where two CMABs are related via a causal abstraction. The authors introduce measures to quantify the quality of the abstraction, such as interventional consistency error and reward discrepancy.
The paper then studies three representative scenarios for transferring information in CAMABs:
Transferring the optimal action from the base CMAB to the abstracted CMAB. The authors show that this is not guaranteed to work, even with an exact abstraction.
Transferring the actions taken in the base CMAB to the abstracted CMAB. The authors analyze the trade-off between confidence and regret in this approach.
Transferring the expected values of rewards from the base CMAB to the abstracted CMAB. The authors propose an algorithm (TExp) that uses this approach and analyze its regret.
Finally, the authors showcase their algorithms on a realistic online advertising scenario and discuss connections to related work.