Interpretable decision making frameworks allow us to easily endow agents with specific goals, risk tolerances, and understanding. Existing decision making systems either forgo interpretability, or pay for it with severely reduced efficiency and large memory requirements. In this paper, we outline DeepID, a neural network approximation of Influence Diagrams, that avoids both pitfalls. We demonstrate how the framework allows for the introduction of robustness in a very transparent and interpretable manner, without increasing the complexity class of the decision problem.
CITATION STYLE
Cooper, H. J., Iyengar, G., & Lin, C. Y. (2019). Deep Influence Diagrams: An Interpretable and Robust Decision Support System. In Lecture Notes in Business Information Processing (Vol. 353, pp. 450–462). Springer Verlag. https://doi.org/10.1007/978-3-030-20485-3_35
Mendeley helps you to discover research relevant for your work.