Logo: to the web site of Uppsala University

uu.sePublications from Uppsala University
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
CalibrationAnalysis.jl: Calibration analysis of probabilistic models in Julia
Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Systems and Control. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Artificial Intelligence.ORCID iD: 0000-0001-9282-053x
Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Division of Systems and Control. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Automatic control. Uppsala University, Disciplinary Domain of Science and Technology, Mathematics and Computer Science, Department of Information Technology, Artificial Intelligence.ORCID iD: 0000-0002-6698-0166
Linköping Univ, Div Stat & Machine Learning, Linköping, Sweden.ORCID iD: 0000-0003-3749-5820
(English)Manuscript (preprint) (Other academic)
Abstract [en]

A probabilistic predictive model tries to capture the uncertainty in its predictions by returning probability distributions of predictions rather than mere point estimates. In safety-critical applications it is particularly important that the uncertainty predicted by the model corresponds to empirically observed uncertainties. Such models, whose predictions are consistent with empirical observations, are called calibrated or reliable. In this article, we present CalibrationAnalysis.jl, a Julia package that can be used to analyze if a probabilistic model is calibrated. Main features of CalibrationAnalysis.jl are the recently proposed kernel calibration error and a set of hypothesis tests of calibration.

Keywords [en]
calibration, probabilistic, modeling, Julia
National Category
Probability Theory and Statistics
Identifiers
URN: urn:nbn:se:uu:diva-500567OAI: oai:DiVA.org:uu-500567DiVA, id: diva2:1751957
Available from: 2023-04-20 Created: 2023-04-20 Last updated: 2023-06-24Bibliographically approved
In thesis
1. Reliable Uncertainty Quantification in Statistical Learning
Open this publication in new window or tab >>Reliable Uncertainty Quantification in Statistical Learning
2023 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Mathematical models are powerful yet simplified abstractions used to study, explain, and predict the behavior of systems of interest. This thesis is concerned with their latter application as predictive models. Predictions of such models are often inherently uncertain, as exemplified in weather forecasting and experienced with epidemiological models during the COVID-19 pandemic. Missing information, such as incomplete atmospheric data, and the very nature of models as approximations ("all models are wrong") imply that predictions are at most approximately correct.

Probabilistic models alleviate this issue by reporting not a single point prediction ("rain"/"no rain") but a probability distribution of all possible outcomes ("80% probability of rain"), representing the uncertainty of a prediction, with the intention to be able to mark predictions as more or less trustworthy. However, simply reporting a probabilistic prediction does not guarantee that the uncertainty estimates are reliable. Calibrated models ensure that the uncertainty expressed by the predictions is consistent with the prediction task and hence the predictions are neither under- nor overconfident. Calibration is important in particular in safety-critical applications such as medical diagnostics and autonomous driving where it is crucial to be able to distinguish between uncertain and trustworthy predictions. Mathematical models do not necessarily possess this property, and in particular complex machine learning models are susceptible to reporting overconfident predictions.

The main contribution of this thesis are new statistical methods for analyzing the calibration of a model, consisting of calibration measures, their estimators, and statistical hypothesis tests based on them. These methods are presented in the five scientific papers in the second part of the thesis. In the first part the reader is introduced to probabilistic predictive models, the analysis of calibration, and positive definite kernels that form the basis of the proposed calibration measures. The contributed tools for calibration analysis cover in principle any predictive model and are applied specifically to classification models, with an arbitrary number of classes, models for regression problems, and models arising from Bayesian inference. This generality is motivated by the need for more detailed calibration analysis of increasingly complex models nowadays. To simplify the use of the statistical methods, a collection of software packages for calibration analysis written in the Julia programming language is made publicly available and supplemented with interfaces to the Python and R programming languages.

Place, publisher, year, edition, pages
Uppsala: Acta Universitatis Upsaliensis, 2023. p. 110
Series
Digital Comprehensive Summaries of Uppsala Dissertations from the Faculty of Science and Technology, ISSN 1651-6214 ; 2275
Keywords
Reliability, Calibration, Uncertainty, Probabilistic Model, Prediction, Julia
National Category
Probability Theory and Statistics
Research subject
Machine learning
Identifiers
urn:nbn:se:uu:diva-500736 (URN)978-91-513-1823-3 (ISBN)
Public defence
2023-06-14, Häggsalen (10132), Ångströmlaboratoriet, Uppsala, 09:15 (English)
Opponent
Supervisors
Available from: 2023-05-23 Created: 2023-04-26 Last updated: 2023-05-23

Open Access in DiVA

No full text in DiVA

Authority records

Widmann, DavidZachariah, Dave

Search in DiVA

By author/editor
Widmann, DavidZachariah, DaveLindsten, Fredrik
By organisation
Division of Systems and ControlArtificial IntelligenceAutomatic control
Probability Theory and Statistics

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 118 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf