Abstract
Methods for learning optimal policies use causal machine learning models to create human-interpretable rules for making choices around the allocation of different policy interventions. However, in realistic policy-making contexts, decision-makers often care about trade-offs between outcomes, not just single-mindedly maximising utility for one outcome. This paper proposes an approach termed Multi-Objective Policy Learning (MOPoL) which combines optimal decision trees for policy learning with a multi-objective Bayesian optimisation approach to explore the trade-off between multiple outcomes. It does this by building a Pareto frontier of non-dominated models for different hyperparameter settings which govern outcome weighting. The method is applied to a real-world case-study of pricing targetting subsididies for anti-malarial medication in Kenya.
Author supplied keywords
Cite
CITATION STYLE
Rehill, P., & Biddle, N. (2025). Policy Learning for Many Outcomes of Interest: Combining Optimal Policy Trees with Multi-objective Bayesian Optimisation. Computational Economics, 66(2), 971–1001. https://doi.org/10.1007/s10614-024-10722-1
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.