deepgp (Q84903)

From MaRDI portal
Bayesian Deep Gaussian Processes using MCMC
Language Label Description Also known as
English
deepgp
Bayesian Deep Gaussian Processes using MCMC

    Statements

    0 references
    1.1.0
    15 December 2022
    0 references
    0.1.0
    29 October 2020
    0 references
    0.2.0
    16 December 2020
    0 references
    0.2.1
    15 July 2021
    0 references
    0.3.0
    24 November 2021
    0 references
    0.3.1
    7 December 2021
    0 references
    1.0.0
    8 April 2022
    0 references
    1.0.1
    20 June 2022
    0 references
    1.1.1
    7 August 2023
    0 references
    0 references
    0 references
    7 August 2023
    0 references
    Performs Bayesian posterior inference for deep Gaussian processes following Sauer, Gramacy, and Higdon (2023, <arXiv:2012.08015>). See Sauer (2023, <http://hdl.handle.net/10919/114845>) for comprehensive methodological details and <https://bitbucket.org/gramacylab/deepgp-ex/> for a variety of coding examples. Models are trained through MCMC including elliptical slice sampling of latent Gaussian layers and Metropolis-Hastings sampling of kernel hyperparameters. Vecchia-approximation for faster computation is implemented following Sauer, Cooper, and Gramacy (2022, <arXiv:2204.02904>). Downstream tasks include sequential design through active learning Cohn/integrated mean squared error (ALC/IMSE; Sauer, Gramacy, and Higdon, 2023), optimization through expected improvement (EI; Gramacy, Sauer, and Wycoff, 2021 <arXiv:2112.07457>), and contour location through entropy (Sauer, 2023). Models extend up to three layers deep; a one layer model is equivalent to typical Gaussian process regression. Incorporates OpenMP and SNOW parallelization and utilizes C/C++ under the hood.
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references