kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Non-Parametric Calibration for Classification
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL. Univ Tubingen, Tubingen, Germany.;Tech Univ Munich, Munich, Germany..ORCID iD: 0000-0003-2261-1331
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.ORCID iD: 0000-0002-5750-9655
Tech Univ Munich, German Aerosp Ctr DLR, Munich, Germany..
2020 (English)In: International Conference on Artificial Intelligence and Statistics, Vol 108 / [ed] Chiappa, S Calandra, R, ML Research Press , 2020Conference paper, Published paper (Refereed)
Abstract [en]

Many applications of classification methods not only require high accuracy but also reliable estimation of predictive uncertainty. However, while many current classification frameworks, in particular deep neural networks, achieve high accuracy, they tend to incorrectly estimate uncertainty. In this paper, we propose a method that adjusts the confidence estimates of a general classifier such that they approach the probability of classifying correctly. In contrast to existing approaches, our calibration method employs a non-parametric representation using a latent Gaussian process, and is specifically designed for multi-class classification. It can be applied to any classifier that outputs confidence estimates and is not limited to neural networks. We also provide a theoretical analysis regarding the over- and underconfidence of a classifier and its relationship to calibration, as well as an empirical outlook for calibrated active learning. In experiments we show the universally strong performance of our method across different classifiers and benchmark data sets, in particular for state-of-the art neural network architectures.

Place, publisher, year, edition, pages
ML Research Press , 2020.
Series
Proceedings of Machine Learning Research, ISSN 2640-3498 ; 108
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:kth:diva-285711ISI: 000559931303078Scopus ID: 2-s2.0-85161931704OAI: oai:DiVA.org:kth-285711DiVA, id: diva2:1536317
Conference
23rd International Conference on Artificial Intelligence and Statistics (AISTATS), AUG 26-28, 2020, ELECTR NETWORK
Note

QC 20210720

Available from: 2021-03-10 Created: 2021-03-10 Last updated: 2023-11-06Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

ScopusFulltext

Authority records

Wenger, JonathanKjellström, Hedvig

Search in DiVA

By author/editor
Wenger, JonathanKjellström, Hedvig
By organisation
Robotics, Perception and Learning, RPL
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 49 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf