Understanding and improving pre-trained neural networks via post-hoc analysis사후 분석을 통한 사전 학습된 신경망의 이해와 개선

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 2
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisor양은호-
dc.contributor.authorKim, Sung-Yub-
dc.contributor.author김성엽-
dc.date.accessioned2024-08-08T19:30:58Z-
dc.date.available2024-08-08T19:30:58Z-
dc.date.issued2024-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1098135&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/321980-
dc.description학위논문(박사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[vii, 74 p. :]-
dc.description.abstractDue to the increasing prevalence of foundation models, the need for precisely understanding a pre-trained neural network and its predictions has become paramount. This paper aims to improve post-analysis methods for pre-trained neural networks. Among various post-analysis methods, this dissertation identifies and solves the limitations of the flat minima hypothesis, which is a parameter-based post-hoc analysis, and the influence function, which is a sample-based post-analysis. First, this dissertation demonstrates that the Laplace Approximation in Bayesian deep learning is susceptible to reparameterization, similar to the flat minima hypothesis. Then, we solve this problem through a new concept called connectivity. Second, this dissertation shows the additivity of the Graph Influence Function is broken for neighboring edges. Based on this, we propose an efficient algorithm that removes noisy edges that deteriorate the generalization of graph neural networks. Lastly, this dissertation shows the unimodal distributional bias of self-influence caused by the bilinearity of the influence function. Then, we propose a novel non-linear influence approximation based on Geometric Ensemble to mitigate this distributional bias.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subject기계학습▼a심층학습▼a일반화▼a최적화-
dc.subjectMachine learning▼aDeep learning▼aGeneralization▼aOptimization-
dc.titleUnderstanding and improving pre-trained neural networks via post-hoc analysis-
dc.title.alternative사후 분석을 통한 사전 학습된 신경망의 이해와 개선-
dc.typeThesis(Ph.D)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :김재철AI대학원,-
dc.contributor.alternativeauthorYang, Eunho-
Appears in Collection
AI-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0