Understanding and improving pre-trained neural networks via post-hoc analysis사후 분석을 통한 사전 학습된 신경망의 이해와 개선

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 1
  • Download : 0
Due to the increasing prevalence of foundation models, the need for precisely understanding a pre-trained neural network and its predictions has become paramount. This paper aims to improve post-analysis methods for pre-trained neural networks. Among various post-analysis methods, this dissertation identifies and solves the limitations of the flat minima hypothesis, which is a parameter-based post-hoc analysis, and the influence function, which is a sample-based post-analysis. First, this dissertation demonstrates that the Laplace Approximation in Bayesian deep learning is susceptible to reparameterization, similar to the flat minima hypothesis. Then, we solve this problem through a new concept called connectivity. Second, this dissertation shows the additivity of the Graph Influence Function is broken for neighboring edges. Based on this, we propose an efficient algorithm that removes noisy edges that deteriorate the generalization of graph neural networks. Lastly, this dissertation shows the unimodal distributional bias of self-influence caused by the bilinearity of the influence function. Then, we propose a novel non-linear influence approximation based on Geometric Ensemble to mitigate this distributional bias.
Advisors
양은호researcher
Description
한국과학기술원 :김재철AI대학원,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(박사) - 한국과학기술원 : 김재철AI대학원, 2024.2,[vii, 74 p. :]

Keywords

기계학습▼a심층학습▼a일반화▼a최적화; Machine learning▼aDeep learning▼aGeneralization▼aOptimization

URI
http://hdl.handle.net/10203/321980
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1098135&flag=dissertation
Appears in Collection
AI-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0