Carnegie Mellon University
Browse

Uncertainty Quantification in Simulation-based Inference

Download (16.68 MB)
thesis
posted on 2021-08-09, 19:07 authored by Niccolo DalmassoNiccolo Dalmasso
Complex phenomena in engineering and the sciences are often modeled by feed-forward simulators that implicitly encode the likelihood function. Classical methods for constructing confidence sets and hypothesis tests are poorly suited for such settings. While the field of simulation-based inference has undergone a revolution in terms of the complexity of problems that can be tackled, the development on the statistical methodology front has fallen behind. Indeed, many techniques have been developed for learning a surrogate likelihood using forward-simulated data, but these methods do not guarantee frequentist confidence sets and tests with nominal coverage and Type I error control, respectively, outside the asymptotic and low-dimensional regimes. In this thesis we introduce a series of statistical tools for uncertainty quantification in a simulation-based inference setting. In the first part of the thesis, we provide inferential tools with frequentist guarantees for a high-dimensional simulator-based setting. We introduce a statistical framework that unifies classical statistics with modern machine learning algorithms to achieve the following
goals: (i) confidence sets and hypothesis tests with finite-sample guarantees of coverage and power, (ii) diagnostics for checking empirical coverage over the entire parameter space and (iii) scalable and modular procedure which can also be used with other simulation-based approaches. We showcase the applicability of this framework across a diverse range of data and parameter settings. In the second part of the thesis, we consider the problem of assessing the quality of fit of approximate likelihood models. Approximate likelihood models are used in settings with high-resolution and computationally intensive simulators for faster inference. We propose a statistically consistent validation method that can pinpoint the locations in the parameter space where the fit is inadequate as well as provide insights as to how the high-resolution simulator and approximate likelihood model differ In the third and final part of the thesis, we propose a new flexible method for nonparametric conditional density estimation that can transform any neural network
regression architecture into a conditional density estimator. We demonstrate the versatility of our approach for two applications with convolutional and recurrent neural networks, respectively.

History

Date

2021-04-21

Degree Type

  • Dissertation

Department

  • Statistics

Degree Name

  • Doctor of Philosophy (PhD)

Advisor(s)

Ann Lee

Usage metrics

    Licence

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC