Markov Information Bottleneck to Improve Information Flow in Stochastic Neural Networks

Cited 4 time in webofscience Cited 3 time in scopus
  • Hit : 392
  • Download : 165
DC FieldValueLanguage
dc.contributor.authorThanh Tang Nguyenko
dc.contributor.authorChoi, Jaesikko
dc.date.accessioned2019-12-13T07:25:53Z-
dc.date.available2019-12-13T07:25:53Z-
dc.date.created2019-12-02-
dc.date.created2019-12-02-
dc.date.created2019-12-02-
dc.date.created2019-12-02-
dc.date.issued2019-10-
dc.identifier.citationENTROPY, v.21, no.10-
dc.identifier.issn1099-4300-
dc.identifier.urihttp://hdl.handle.net/10203/268921-
dc.description.abstractWhile rate distortion theory compresses data under a distortion constraint, information bottleneck (IB) generalizes rate distortion theory to learning problems by replacing a distortion constraint with a constraint of relevant information. In this work, we further extend IB to multiple Markov bottlenecks (i.e., latent variables that form a Markov chain), namely Markov information bottleneck (MIB), which particularly fits better in the context of stochastic neural networks (SNNs) than the original IB. We show that Markov bottlenecks cannot simultaneously achieve their information optimality in a non-collapse MIB, and thus devise an optimality compromise. With MIB, we take the novel perspective that each layer of an SNN is a bottleneck whose learning goal is to encode relevant information in a compressed form from the data. The inference from a hidden layer to the output layer is then interpreted as a variational approximation to the layer's decoding of relevant information in the MIB. As a consequence of this perspective, the maximum likelihood estimate (MLE) principle in the context of SNNs becomes a special case of the variational MIB. We show that, compared to MLE, the variational MIB can encourage better information flow in SNNs in both principle and practice, and empirically improve performance in classification, adversarial robustness, and multi-modal learning in MNIST.-
dc.languageEnglish-
dc.publisherMDPI-
dc.titleMarkov Information Bottleneck to Improve Information Flow in Stochastic Neural Networks-
dc.typeArticle-
dc.identifier.wosid000495094000058-
dc.identifier.scopusid2-s2.0-85074007561-
dc.type.rimsART-
dc.citation.volume21-
dc.citation.issue10-
dc.citation.publicationnameENTROPY-
dc.identifier.doi10.3390/e21100976-
dc.contributor.localauthorChoi, Jaesik-
dc.contributor.nonIdAuthorThanh Tang Nguyen-
dc.description.isOpenAccessY-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorinformation bottleneck-
dc.subject.keywordAuthorstochastic neural networks-
dc.subject.keywordAuthorvariational inference-
dc.subject.keywordAuthormachine learning-
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 4 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0