PAC-Bayesian Bounds for Deep Gaussian Processes
From MaRDI portal
Publication:6325752
arXiv1909.09985MaRDI QIDQ6325752FDOQ6325752
Publication date: 22 September 2019
Abstract: Variational approximation techniques and inference for stochastic models in machine learning has gained much attention the last years. Especially in the case of Gaussian Processes (GP) and their deep versions, Deep Gaussian Processes (DGPs), these viewpoints improved state of the art work. In this paper we introduce Probably Approximately Correct (PAC)-Bayesian risk bounds for DGPs making use of variational approximations. We show that the minimization of PAC-Bayesian generalization risk bounds maximizes the variational lower bounds belonging to the specific DGP model. We generalize the loss function property of the log likelihood loss function in the context of PAC-Bayesian risk bounds to the quadratic-form-Gaussian case. Consistency results are given and an oracle-type inequality gives insights in the convergence between the raw model (predictor without variational approximation) and our variational models (predictor for the variational approximation). Furthermore, we give extensions of our main theorems for specific assumptions and parameter cases. Moreover, we show experimentally the evolution of the consistency results for two Deep Recurrent Gaussian Processes (DRGP) modeling time-series, namely the recurrent Gaussian Process (RGP) and the DRGP with Variational Sparse Spectrum approximation, namely DRGP-(V)SS.
This page was built for publication: PAC-Bayesian Bounds for Deep Gaussian Processes
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6325752)