Data association and feature selection with Gaussian processes가우시안 확률과정을 이용한 자료연관 기법과 특성 선택 기법

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 143
  • Download : 0
DC FieldValueLanguage
dc.contributor.advisorHwang, Ganguk-
dc.contributor.advisor황강욱-
dc.contributor.authorJeon, Younghwan-
dc.date.accessioned2023-06-22T19:33:54Z-
dc.date.available2023-06-22T19:33:54Z-
dc.date.issued2022-
dc.identifier.urihttp://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=1007827&flag=dissertationen_US
dc.identifier.urihttp://hdl.handle.net/10203/308576-
dc.description학위논문(박사) - 한국과학기술원 : 수리과학과, 2022.8,[v, 54 p. :]-
dc.description.abstractGaussian process (GP) has been widely used in supervised learning as a prior process to directly estimate a hidden and complicated relation between input and output. Due to this ability of GP, it is recently applied in several areas of data analysis and machine learning, including data association and feature selection. In this proposal we propose new methods with GPs for data association and feature selection. As the goal of data association is separating observations from different sources, we propose a Bayesian approach based on a mixture of GPs having two key components, the assignment probabilities and the GPs. In the proposed approach, the two key components are simultaneously updated according to observations through an efficient Expectation-Maximization (EM) algorithm that we newly develop. The proposed approach is thus more adaptive to the observations than the existing GP based approaches. We also provide a theoretical analysis to show the effectiveness of the Bayesian update in the proposed approach. Next, we examine a recently proposed feature selection method that measures the feature relevance by estimating dissimilarity between the predictive distributions of GP at a training sample and its perturbed sample by a small amount. However, this existing method with GP suffers from the scalability problem and hence needs refinement for its applicability to large data sets. Moreover, it uses the Kullback-Leibler (KL) divergence in sensitivity analysis for feature selection, but we theoretically show that the KL divergence under-estimates the relevance of important features in some cases of classification. Hence we propose a new method with GP to remedy such limitations of the existing method for better feature selection. Throughout experiments with synthetic and real data sets we show that the proposed methods outperform the existing methods.-
dc.languageeng-
dc.publisher한국과학기술원-
dc.subjectGaussian process▼adata association▼afeature selection▼aBayesian model▼avariational inference-
dc.subject가우시안 확률과정▼a자료연관 기법▼a특성 선택 기법▼a베이지안 모델▼a변분 추론-
dc.titleData association and feature selection with Gaussian processes-
dc.title.alternative가우시안 확률과정을 이용한 자료연관 기법과 특성 선택 기법-
dc.typeThesis(Ph.D)-
dc.identifier.CNRN325007-
dc.description.department한국과학기술원 :수리과학과,-
dc.contributor.alternativeauthor전영환-
Appears in Collection
MA-Theses_Ph.D.(박사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0