Training neural network through learning the derivative of loss function손실 함수의 도함수의 학습을 통한 인공신경망 학습

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 71
  • Download : 0
Recent studies and applications in machine learning shows that neural network using gradient descent-based methods suits the need in many cases. Especially, there have been many researches which talks about the importance of choosing a loss function, when using such methods. While there also have been studies which accepted the concept of meta-learning and substituted the loss function with a computational graph or a neural network, they didn’t question about the necessity of the loss function if they were to be replaced. This paper first looks the structural properties of neural networks. Also, the paper rethinks the role of loss function in neural networks, then gives an idea and method of learning the derivative of the loss function, and finally shows some experimental results on regression.
Advisors
최성희researcher
Description
한국과학기술원 :전산학부,
Publisher
한국과학기술원
Issue Date
2024
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 전산학부, 2024.2,[iii, 15 p. :]

Keywords

메타 학습▼a최적화▼a경사하강법▼a딥러닝▼a인공 신경망; Meta learning▼aOptimization▼aGradient descent▼aDeep learning▼aNeural network

URI
http://hdl.handle.net/10203/321789
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1097314&flag=dissertation
Appears in Collection
CS-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0