Interpreting and Explaining Deep Neural Networks: A Perspective on Time Series Data

5Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Explainable and interpretable machine learning models and algorithms are important topics which have received growing attention from research, application and administration. Many complex Deep Neural Networks (DNNs) are often perceived as black-boxes. Researchers would like to be able to interpret what the DNN has learned in order to identify biases and failure models and improve models. In this tutorial, we will provide a comprehensive overview on methods to analyze deep neural networks and an insight how those interpretable and explainable methods help us understand time series data.

Cite

CITATION STYLE

APA

Choi, J. (2020). Interpreting and Explaining Deep Neural Networks: A Perspective on Time Series Data. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 3563–3564). Association for Computing Machinery. https://doi.org/10.1145/3394486.3406478

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free