Abstract
Explainable and interpretable machine learning models and algorithms are important topics which have received growing attention from research, application and administration. Many complex Deep Neural Networks (DNNs) are often perceived as black-boxes. Researchers would like to be able to interpret what the DNN has learned in order to identify biases and failure models and improve models. In this tutorial, we will provide a comprehensive overview on methods to analyze deep neural networks and an insight how those interpretable and explainable methods help us understand time series data.
Author supplied keywords
Cite
CITATION STYLE
Choi, J. (2020). Interpreting and Explaining Deep Neural Networks: A Perspective on Time Series Data. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 3563–3564). Association for Computing Machinery. https://doi.org/10.1145/3394486.3406478
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.