ESD: Expected Squared Difference as a Tuning-Free Trainable Calibration Measure

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 74
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorYoon, Hee Sukko
dc.contributor.authorTee, Joshua Tian Jinko
dc.contributor.authorYoon, Eunseopko
dc.contributor.authorYoon, Sunjaeko
dc.contributor.authorKim, Gwangsuko
dc.contributor.authorLi, Yingzhenko
dc.contributor.authorYoo, Chang-Dongko
dc.date.accessioned2023-06-13T01:00:52Z-
dc.date.available2023-06-13T01:00:52Z-
dc.date.created2023-06-13-
dc.date.created2023-06-13-
dc.date.issued2023-05-02-
dc.identifier.citationInternational Conference on Learning Representations (ICLR) 2023-
dc.identifier.urihttp://hdl.handle.net/10203/307224-
dc.description.abstractStudies have shown that modern neural networks tend to be poorly calibrated due to over-confident predictions. Traditionally, post-processing methods have been used to calibrate the model after training. In recent years, various trainable calibration measures have been proposed to incorporate them directly into the training process. However, these methods all incorporate internal hyperparameters, and the performance of these calibration objectives relies on tuning these hyperparameters, incurring more computational costs as the size of neural networks and datasets become larger. As such, we present Expected Squared Difference (ESD), a tuning-free (i.e., hyperparameter-free) trainable calibration objective loss, where we view the calibration error from the perspective of the squared difference between the two expectations. With extensive experiments on several architectures (CNNs, Transformers) and datasets, we demonstrate that (1) incorporating ESD into the training improves model calibration in various batch size settings without the need for internal hyperparameter tuning, (2) ESD yields the best-calibrated results compared with previous approaches, and (3) ESD drastically improves the computational costs required for calibration during training due to the absence of internal hyperparameter. The code is publicly accessible at https://github.com/hee-suk-yoon/ESD.-
dc.languageEnglish-
dc.publisherInternational Conference on Learning Representations-
dc.titleESD: Expected Squared Difference as a Tuning-Free Trainable Calibration Measure-
dc.typeConference-
dc.type.rimsCONF-
dc.citation.publicationnameInternational Conference on Learning Representations (ICLR) 2023-
dc.identifier.conferencecountryRW-
dc.identifier.conferencelocationKigali-
dc.contributor.localauthorYoo, Chang-Dong-
dc.contributor.nonIdAuthorTee, Joshua Tian Jin-
dc.contributor.nonIdAuthorYoon, Eunseop-
dc.contributor.nonIdAuthorLi, Yingzhen-
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0