DC Field | Value | Language |
---|---|---|
dc.contributor.author | Choi, Jaesik | ko |
dc.date.accessioned | 2021-07-07T01:10:15Z | - |
dc.date.available | 2021-07-07T01:10:15Z | - |
dc.date.created | 2021-07-07 | - |
dc.date.issued | 2020-08-23 | - |
dc.identifier.citation | 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 2020, pp.3563 - 3564 | - |
dc.identifier.uri | http://hdl.handle.net/10203/286463 | - |
dc.description.abstract | Explainable and interpretable machine learning models and algorithms are important topics which have received growing attention from research, application and administration. Many complex Deep Neural Networks (DNNs) are often perceived as black-boxes. Researchers would like to be able to interpret what the DNN has learned in order to identify biases and failure models and improve models. In this tutorial, we will provide a comprehensive overview on methods to analyze deep neural networks and an insight how those interpretable and explainable methods help us understand time series data. | - |
dc.language | English | - |
dc.publisher | Association for Computing Machinery | - |
dc.title | Interpreting and Explaining Deep Neural Networks: A Perspective on Time Series Data | - |
dc.type | Conference | - |
dc.identifier.scopusid | 2-s2.0-85090401893 | - |
dc.type.rims | CONF | - |
dc.citation.beginningpage | 3563 | - |
dc.citation.endingpage | 3564 | - |
dc.citation.publicationname | 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 2020 | - |
dc.identifier.conferencecountry | US | - |
dc.identifier.conferencelocation | Virtual | - |
dc.identifier.doi | 10.1145/3394486.3406478 | - |
dc.contributor.localauthor | Choi, Jaesik | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.