Predictive, finite-sample model choice for time series under stationarity and non-stationarity (Q143634): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
Normalize DOI.
 
(5 intermediate revisions by 4 users not shown)
aliases / en / 0aliases / en / 0
 
Predictive, finite-sample model choice for time series under stationarity and non-stationarity
description / endescription / en
scientific article from arXiv
scientific article; zbMATH DE number 7113729
Property / DOI
 
Property / DOI: 10.48550/arXiv.1611.04460 / rank
Normal rank
 
Property / publication date
 
4 October 2019
Timestamp+2019-10-04T00:00:00Z
Timezone+00:00
CalendarGregorian
Precision1 day
Before0
After0
Property / publication date: 4 October 2019 / rank
 
Normal rank
Property / author
 
Property / author: Tobias Kley / rank
 
Normal rank
Property / author
 
Property / author: Piotr Fryzlewicz / rank
 
Normal rank
Property / author
 
Property / author: Philip Preuss / rank
 
Normal rank
Property / DOI
 
Property / DOI: 10.1214/19-EJS1606 / rank
 
Normal rank
Property / DOI
 
Property / DOI: 10.48550/ARXIV.1611.04460 / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / title
 
Predictive, finite-sample model choice for time series under stationarity and non-stationarity (English)
Property / title: Predictive, finite-sample model choice for time series under stationarity and non-stationarity (English) / rank
 
Normal rank
Property / zbMATH Open document ID
 
Property / zbMATH Open document ID: 1432.62323 / rank
 
Normal rank
Property / published in
 
Property / published in: Electronic Journal of Statistics / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://arxiv.org/abs/1611.04460 / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://projecteuclid.org/euclid.ejs/1569895286 / rank
 
Normal rank
Property / review text
 
The paper is concerned with the bias-variance trade-off problem in the context of stationarity versus non-stationarity modeling for prediction. The authors find that a simple stationary model can often perform better in terms of finite sample prediction than a more complicated non-stationary one. The authors illustrate that phenomenon first at a time varying AR(2) process, where they compare the prediction performance with that of a time varying AR(1) process and with stationary AR(1) and AR(2) processes, with the findings that for not too large sample sizes, the stationary models may very well give more accurate prediction in terms of the empirical mean squared errors. The authors then continue to develop a general methodology to find a model for prediction of finite sample data. For that, they divide the observation sets into three sets, namely the training set \(M_0\), a validation set \(M_1\) and a final validation set \(M_2\). Based on the observations in \(M_0\), they calculate the locally stationary model that predicts best (among certain other locally stationary models) into the set \(M_1\) in terms of minimised empirical mean square prediction errors for linear \(h\)-step predictors, and similarly the best stationary model with this property. Whether to choose the best locally stationary model or the best stationary model is then decided by the comparison of the performance of both models on the second evaluation set \(M_2\). The authors then show rigorously under certain assumptions that with high probability, the chosen of the two models (best locally stationary / best stationary) will perform empirically better in forecasting the future of not yet observed values. Various simulations are presented and an \texttt{R} package is provided. The authors apply their method to three real data sets, namely to London housing prices, to temperatures and to the volatility around the time of the EU referendum in the UK in 2016. As a further theoretical result, they prove that the localised Yule-Walker estimator in locally stationary models is strongly, uniformly consistent.
Property / review text: The paper is concerned with the bias-variance trade-off problem in the context of stationarity versus non-stationarity modeling for prediction. The authors find that a simple stationary model can often perform better in terms of finite sample prediction than a more complicated non-stationary one. The authors illustrate that phenomenon first at a time varying AR(2) process, where they compare the prediction performance with that of a time varying AR(1) process and with stationary AR(1) and AR(2) processes, with the findings that for not too large sample sizes, the stationary models may very well give more accurate prediction in terms of the empirical mean squared errors. The authors then continue to develop a general methodology to find a model for prediction of finite sample data. For that, they divide the observation sets into three sets, namely the training set \(M_0\), a validation set \(M_1\) and a final validation set \(M_2\). Based on the observations in \(M_0\), they calculate the locally stationary model that predicts best (among certain other locally stationary models) into the set \(M_1\) in terms of minimised empirical mean square prediction errors for linear \(h\)-step predictors, and similarly the best stationary model with this property. Whether to choose the best locally stationary model or the best stationary model is then decided by the comparison of the performance of both models on the second evaluation set \(M_2\). The authors then show rigorously under certain assumptions that with high probability, the chosen of the two models (best locally stationary / best stationary) will perform empirically better in forecasting the future of not yet observed values. Various simulations are presented and an \texttt{R} package is provided. The authors apply their method to three real data sets, namely to London housing prices, to temperatures and to the volatility around the time of the EU referendum in the UK in 2016. As a further theoretical result, they prove that the localised Yule-Walker estimator in locally stationary models is strongly, uniformly consistent. / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 62M20 / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 62M10 / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 62P20 / rank
 
Normal rank
Property / zbMATH DE Number
 
Property / zbMATH DE Number: 7113729 / rank
 
Normal rank
Property / zbMATH Keywords
 
bias-variance trade-off
Property / zbMATH Keywords: bias-variance trade-off / rank
 
Normal rank
Property / zbMATH Keywords
 
forecasting
Property / zbMATH Keywords: forecasting / rank
 
Normal rank
Property / zbMATH Keywords
 
prediction
Property / zbMATH Keywords: prediction / rank
 
Normal rank
Property / zbMATH Keywords
 
Yule-Walker estimate
Property / zbMATH Keywords: Yule-Walker estimate / rank
 
Normal rank
Property / zbMATH Keywords
 
local stationarity versus stationarity
Property / zbMATH Keywords: local stationarity versus stationarity / rank
 
Normal rank
Property / describes a project that uses
 
Property / describes a project that uses: R / rank
 
Normal rank
Property / cites work
 
Property / cites work: Fitting autoregressive models for prediction / rank
 
Normal rank
Property / cites work
 
Property / cites work: A fundamental relation between predictor identification and power spectrum estimation / rank
 
Normal rank
Property / cites work
 
Property / cites work: Asymptotic prediction mean squared error for vector autoregressive models / rank
 
Normal rank
Property / cites work
 
Property / cites work: On large deviations in the Gaussian autoregressive process: Stable, unstable and explosive cases / rank
 
Normal rank
Property / cites work
 
Property / cites work: Sharp large deviations for Gaussian quadratic forms with applications / rank
 
Normal rank
Property / cites work
 
Property / cites work: Large deviations for quadratic forms of stationary Gaussian processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Exponential inequalities for self-normalized martingales with applications / rank
 
Normal rank
Property / cites work
 
Property / cites work: Asymptotically efficient autoregressive model selection for multistep prediction / rank
 
Normal rank
Property / cites work
 
Property / cites work: Quantile Spectral Analysis for Locally Stationary Time Series / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4082810 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Time series: theory and methods. / rank
 
Normal rank
Property / cites work
 
Property / cites work: Asymptotic inference for nearly nonstationary AR(1) processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Localized Realized Volatility Modeling / rank
 
Normal rank
Property / cites work
 
Property / cites work: PREDICTION‐FOCUSED MODEL SELECTION FOR AUTOREGRESSIVE MODELS / rank
 
Normal rank
Property / cites work
 
Property / cites work: The Focused Information Criterion / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the Kullback-Leibler information divergence of locally stationary processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Fitting time series models to nonstationary processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Time series analysis: Methods and applications / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the Optimal Segment Length for Parameter Estimates for Locally Stationary Time Series / rank
 
Normal rank
Property / cites work
 
Property / cites work: Predictive Inference for Locally Stationary Time Series With an Application to Climate Data / rank
 
Normal rank
Property / cites work
 
Property / cites work: A Measure of Stationarity in Locally Stationary Processes With Applications to Testing / rank
 
Normal rank
Property / cites work
 
Property / cites work: A test for second-order stationarity of a time series based on the discrete Fourier transform / rank
 
Normal rank
Property / cites work
 
Property / cites work: Parameter estimation for nearly nonstationary AR(1) processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Mixing properties of ARCH and time-varying ARCH processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Adaptive forecasting in the presence of recent and ongoing structural change / rank
 
Normal rank
Property / cites work
 
Property / cites work: Mixed autoregressive-moving average multivariate processes with time- dependent coefficients / rank
 
Normal rank
Property / cites work
 
Property / cites work: Autocorrelation, autoregression and autoregressive approximation / rank
 
Normal rank
Property / cites work
 
Property / cites work: Simultaneous confidence bands for Yule-Walker estimators and order selection / rank
 
Normal rank
Property / cites work
 
Property / cites work: Simultaneous confidence bands for sequential autoregressive fitting / rank
 
Normal rank
Property / cites work
 
Property / cites work: Predictive, finite-sample model choice for time series under stationarity and non-stationarity / rank
 
Normal rank
Property / cites work
 
Property / cites work: Least squares estimates in stochastic regression models with applications to identification and control of dynamic systems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Nonparametric risk bounds for time-series forecasting / rank
 
Normal rank
Property / cites work
 
Property / cites work: On recursive estimation for time varying autoregressive processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: A Test for Second-Order Stationarity and Approximate Confidence Intervals for Localized Autocovariances for Locally Stationary Time Series / rank
 
Normal rank
Property / cites work
 
Property / cites work: Estimation and Forecasting of Locally Stationary Processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Testing temporal constancy of the spectral structure of a time series / rank
 
Normal rank
Property / cites work
 
Property / cites work: Validating Stationarity Assumptions in Time Series Analysis by Rolling Local Periodograms / rank
 
Normal rank
Property / cites work
 
Property / cites work: Measuring the Advantages of Multivariate vs. Univariate Forecasts / rank
 
Normal rank
Property / cites work
 
Property / cites work: Model-free prediction and regression. A transformation-based approach to inference / rank
 
Normal rank
Property / cites work
 
Property / cites work: Discriminating between long-range dependence and non-stationarity / rank
 
Normal rank
Property / cites work
 
Property / cites work: A test for stationarity based on empirical processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5521155 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3323077 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3902363 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Cross validation for locally stationary processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: ORDER SELECTION IN ARMA MODELS USING THE FOCUSED INFORMATION CRITERION / rank
 
Normal rank
Property / cites work
 
Property / cites work: Prediction of weakly locally stationary processes by auto-regression / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4002143 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5631967 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Nonparametric regression for locally stationary time series / rank
 
Normal rank
Property / cites work
 
Property / cites work: Detecting gradual changes in locally stationary processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: A Wavelet-Based Test for Stationarity / rank
 
Normal rank
Property / cites work
 
Property / cites work: Feature matching in time series modeling / rank
 
Normal rank
Property / cites work
 
Property / cites work: Moderate deviation principle for autoregressive processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Adaptive Order Determination for Constructing Time Series Forecasting Models / rank
 
Normal rank
links / mardi / namelinks / mardi / name
 

Latest revision as of 01:02, 28 December 2024

scientific article; zbMATH DE number 7113729
  • Predictive, finite-sample model choice for time series under stationarity and non-stationarity
Language Label Description Also known as
English
Predictive, finite-sample model choice for time series under stationarity and non-stationarity
scientific article; zbMATH DE number 7113729
  • Predictive, finite-sample model choice for time series under stationarity and non-stationarity

Statements

14 November 2016
0 references
4 October 2019
0 references
math.ST
0 references
stat.ME
0 references
stat.TH
0 references
0 references
0 references
0 references
0 references
0 references
0 references
Predictive, finite-sample model choice for time series under stationarity and non-stationarity (English)
0 references
The paper is concerned with the bias-variance trade-off problem in the context of stationarity versus non-stationarity modeling for prediction. The authors find that a simple stationary model can often perform better in terms of finite sample prediction than a more complicated non-stationary one. The authors illustrate that phenomenon first at a time varying AR(2) process, where they compare the prediction performance with that of a time varying AR(1) process and with stationary AR(1) and AR(2) processes, with the findings that for not too large sample sizes, the stationary models may very well give more accurate prediction in terms of the empirical mean squared errors. The authors then continue to develop a general methodology to find a model for prediction of finite sample data. For that, they divide the observation sets into three sets, namely the training set \(M_0\), a validation set \(M_1\) and a final validation set \(M_2\). Based on the observations in \(M_0\), they calculate the locally stationary model that predicts best (among certain other locally stationary models) into the set \(M_1\) in terms of minimised empirical mean square prediction errors for linear \(h\)-step predictors, and similarly the best stationary model with this property. Whether to choose the best locally stationary model or the best stationary model is then decided by the comparison of the performance of both models on the second evaluation set \(M_2\). The authors then show rigorously under certain assumptions that with high probability, the chosen of the two models (best locally stationary / best stationary) will perform empirically better in forecasting the future of not yet observed values. Various simulations are presented and an \texttt{R} package is provided. The authors apply their method to three real data sets, namely to London housing prices, to temperatures and to the volatility around the time of the EU referendum in the UK in 2016. As a further theoretical result, they prove that the localised Yule-Walker estimator in locally stationary models is strongly, uniformly consistent.
0 references
bias-variance trade-off
0 references
forecasting
0 references
prediction
0 references
Yule-Walker estimate
0 references
local stationarity versus stationarity
0 references
0 references
0 references
0 references
0 references
0 references
0 references

Identifiers

0 references
0 references
0 references
0 references
0 references