• Tidak ada hasil yang ditemukan

07350015%2E2014%2E977445

N/A
N/A
Protected

Academic year: 2017

Membagikan "07350015%2E2014%2E977445"

Copied!
4
0
0

Teks penuh

(1)

Full Terms & Conditions of access and use can be found at

http://www.tandfonline.com/action/journalInformation?journalCode=ubes20

Download by: [Universitas Maritim Raja Ali Haji] Date: 11 January 2016, At: 19:15

Journal of Business & Economic Statistics

ISSN: 0735-0015 (Print) 1537-2707 (Online) Journal homepage: http://www.tandfonline.com/loi/ubes20

Comment

Andrew J. Patton

To cite this article: Andrew J. Patton (2015) Comment, Journal of Business & Economic

Statistics, 33:1, 22-24, DOI: 10.1080/07350015.2014.977445

To link to this article: http://dx.doi.org/10.1080/07350015.2014.977445

Published online: 26 Jan 2015.

Submit your article to this journal

Article views: 106

View related articles

(2)

22 Journal of Business & Economic Statistics, January 2015

Comment

Andrew J. P

ATTON

Department of Economics, Duke University, 213 Social Sciences Building, Box 90097, Durham, NC 27708 ([email protected])

1. INTRODUCTION

This interesting article provides an insider’s view of the early days of the “modern” forecast comparison literature, which might be dated as beginning with Diebold and Mariano (1995) and West (1996), papers that spawned numerous extensions and refinements over the subsequent years.1The article also reviews

how forecast evaluation methods have been employed in the literature, offering some criticism of the use of split-sample forecast comparison techniques (like the Diebold-Mariano test) when applied to model comparison rather than forecast com-parison. My discussion is a personal perspective on a personal perspective on the DM test, and it is perhaps a bold move to debate the DM test with “D,” but below I will do so.

The key ingredient in a DM test is the loss differential:

dt =L

→R+is some loss function, mapping from the

range of possible values for the target variable and the forecast to the nonnegative real line.2With this variable defined, a test of equal predictive accuracy is easily seen to be equivalent to a simple test that the mean of this scalar time series is zero:

H0:E

Diebold praises the virtue of the single assumption that is required to implement the DM test, namely that of covariance stationarity of dt. With such an assumption in place, a

cen-tral limit theorem for the sample mean, ¯dT,can be invoked,

and the properties of the DM test easily established. As noted in his footnote 2, the assumption of covariance stationarity is stronger than needed; what is actually desired is simply a set of assumptions that are sufficient for the DM test statistic to be asymptotically standard Normal under the null hypothesis of equal predictive accuracy. Thus, one might consider swapping Assumption DM (Equation (1) in the article) with the following higher-level assumption:

1I denote this the “modern” forecast comparison literature, as the literature on

forecast evaluation and comparison stretches all the way back to the very birth of econometrics, with an article by Cowles (1933).

2The notation here is slightly more general than that used in Diebold’s article.

This notation allows for loss functions that cannot be expressed in terms of the forecast error, such as those based on proportional errors,y/y,ˆ which commonly appear in volatility forecasting applications, see Patton (2011) for example, as well as those based on sign forecast errors, sgn (y)−sgn ( ˆy), see Cumby and Modest (1987) for example.

In this discussion, I argue that neither covariance stationarity nor asymptotic Normality more generally are needed for a test to be a “DM-type” test.

2. ASYMPTOTIC NORMALITY IS NICE BUT NOT NECESSARY

In his article, Diebold argues that there is an important distinc-tion between comparing forecasts over some historical period and comparing models at their pseudo-true parameter values. I completely concur, and I further suggest that, fundamentally, the biggest distinction between the many tests that have been proposed in the last 20 years is not the shape of the limiting distribution of the test statistics, or the various terms that appear in the asymptotic variance, but rather thenull hypothesis that the statistic is used to test. This hypothesis summarizes the eco-nomic question being posed, and differences can matter greatly for the economic conclusions drawn. This distinction was noted by Diebold and Mariano (1995) and West (1996), but the clearest discussion of this distinction is in Giacomini and White (2006). These latter authors focused explicitly on hypotheses involving estimated parameters, and one of the contributions of their ar-ticle is to provide primitive conditions under which such tests, like the DM test, can be applied.

WCM H0:E

Tests comparing models at their pseudo-true parameter values correspond to “WCM” tests,3while tests comparing forecasts

(from estimated models, or from surveys, judgment forecasts, etc.) correspond to DM-GW tests.

I propose that the defining feature of a “DM-type” test is not the asymptotic Normality of the test statistic (and the simple use of a HAC estimator of the long-run variance ofdt), but rather its

focus on tests offorecasts, rather than tests of models evaluated at their pseudo-true parameter values. If one accepts this as a key feature, then any test of a null like that in Equation (5) might reasonably be called a “DM-type” test, including those that do nothave asymptotically Normal test statistics.

© 2015American Statistical Association Journal of Business & Economic Statistics

January 2015, Vol. 33, No. 1 DOI:10.1080/07350015.2014.977445 3I follow Diebold in calling these tests “WCM” after West (1996) and Clark and

McCracken (2001), although several other authors have focused on tests of nulls involving population parameters, including White (2000), Corradi and Swanson (2002), Elliott, Komunjer, and Timmermann (2005), and Rossi (2005).

(3)

Patton: Comment 23

Table 1. Rejection probabilities under the null

Out-of-sample size

Normal Kiefer-Vogelsang

In-sample size 500 1000 2000 500 1000 2000

500 0.22 0.17 0.17 0.07 0.05 0.03

1000 0.24 0.22 0.20 0.13 0.08 0.05

NOTE: This table presents the proportion of simulations in which the null hypothesis of equal predictive accuracy is rejected at the 0.05 level, in a design in which the null hypothesis is true. All tests are based on a standard Diebold-Mariano test statistic, with HAC standard errors constructed using Newey and West (1987). The “in-sample size” refers to the number of observations used to estimate the forecast models; the “out-of-sample size” (P) refers to the number of observations used to compare the forecast models. In the left panel, the number of lags used in the HAC estimate is set to 3P1/3and critical values are obtained

from the standard Normal distribution. In the right panel, the number of lags is set to

P /2 and critical values are obtained from Kiefer and Vogelsang (2005). Further details are available in Li and Patton (2013).

3. NONNORMAL “DM” TESTS

In a recent working article, Li and Patton (2013) extended the applicability of the tests of Diebold and Mariano (1995), West (1996), White (2000), Giacomini and White (2006), McCracken (2007), and others, to accommodatelatenttarget variables like those encountered in financial applications, for example, fore-casting volatility, correlation, beta, or jump risk. The goal in that article is to provide conditions under which tests based on a proxy constructed using high frequency data (e.g., “real-ized correlation”) have the same properties as the corresponding test based on the true, latent, target variable (e.g., “integrated correlation”). As part of the evaluation of the proposed the-ory, the simple case where the target variableisobservable is considered, which corresponds to a standard DM-type test. One simulation design in Li and Patton (2013) focuses on correlation forecasting, using a bivariate stochastic volatility process from Barndorff-Nielsen and Shephard (2004) as the data-generating process, and comparing forecasts from two DCC models (Engle

2002) for daily correlations.4The null to be tested is

H0:E That is, a “DM-type” null hypothesis. The usual heteroscedas-ticity and autocorrelation (HAC) robustt-statistics for this test were found to be far from Normal under the null hypothesis. After some investigation, it appeared that the need for many lags in the HAC variance estimation was distorting the behavior of the test statistic, prompting the consideration of the “fixed b” asymptotics of Kiefer and Vogelsang (2005). The Kiefer-Vogelsang approach also uses a “t-test,” with possibly many autocorrelations included in the HAC variance estimate, but the limiting distribution is not Normal; rather it follows a nonstan-dard distribution, with critical values provide by Kiefer and Vogelsang. AsTable 1reveals, the Kiefer-Vogelsang approach to obtaining a (non-Normal) limiting distribution for the DMt -statistic provides better size control than the familiar approach based on asymptotic Normality.

4Details on the data-generating process and forecast models are omitted in the

interests of space; interested readers are referred to Li and Patton (2013).

Other examples of non-Normal “DM-type” tests exist in the literature. Giacomini and Rossi (2010) proposed a forecast com-parison test that allows for the possibility of time variation in the relative performance of the competing forecasts over the sample. Importantly, the null hypothesis is a function of the forecasts evaluated using estimated parameters, not at the pa-rameters’ probability limits. The limiting distribution of the test statistic in this application is a functional of a Brownian mo-tion. The “reality check” of White (2000) is presented as a test of a “WCM-type” null, focusing on forecasts generated using the pseudo-true values of the parameters, however, a DM-GW version of the “reality check” is a natural extension (and indeed is hinted at in White’s article). The test statistic in that case has a limiting distribution that is the maximum of a vector of corre-lated Normal random variables, and is thus also non-Normal.

4. SUMMARY

One of the main arguments made in Diebold’s interesting ar-ticle is that if the objective of an analysis is to compare forecasts over some sample period, then the variables of interest are the actual forecasts obtained in that sample period, not those that would be obtained using the pseudo-true parameters of the mod-els (if any) that generated them. I agree, and following through on this point, it suggests that the key aspect of the “DM” ap-proach to forecast comparison is the choice to state the null hypothesis as a function of estimated parameters rather than pseudo-true probability limits. Thus, while covariance station-arity of the loss differences (“Assumption DM” in the article), or, more generally, asymptotic normality of the sample mean of the loss differences, is convenient when applicable, it is not necessary for a “DM-type” test; in some applications a DM-type test will have a non-Normal limit distribution.

ACKNOWLEDGMENTS

I thank Jia Li for helpful comments.

REFERENCES

Barndorff-Nielsen, O. E., and Shephard, N. (2004), “Econometric Analysis of Realized Covariation: High Frequency Based Covariance, Regression, and Correlation in Financial Economics,”Econometrica, 72, 885–925. [23] Clark, T. E., and McCracken, M. W. (2001), “Tests of Equal Forecast Accuracy

and Encompassing for Nested Models,”Journal of Econometrics, 105, 85– 110. [22]

Corradi, V., and Swanson, N. R. (2002), “A Consistent Test for Out of Sample Nonlinear Predictive Ability,”Journal of Econometrics, 110, 353–381. [22] Cowles, A. 3rd (1933), “Can Stock Market Forecasters Forecast?,”

Economet-rica, 1, 309–324. [22]

Cumby, R. E., and Modest, D. M. (1987), “Testing for Market Timing Ability: A Framework for Forecast Evaluation,”Journal of Financial Economics, 19, 169–189. [22]

Diebold, F. X., and Mariano, R. S. (1995), “Comparing Predictive Accuracy,”

Journal of Business and Economic Statistics, 13, 253–263. [22,23] Elliott, G., Komunjer, I., and Timmermann, A. (2005), “Estimation and Testing

of Forecast Rationality under Flexible Loss,”Review of Economic Studies, 72, 1107–1125. [22]

Engle, R. F. (2002), “Dynamic Conditional Correlation: A Simple Class of Mul-tivariate Generalized Autoregressive Conditional Heteroskedasticity Mod-els,”Journal of Business and Economic Statistics, 20, 339–350. [23] Giacomini, R., and Rossi, B. (2010), “Forecast Comparisons in Unstable

Envi-ronments,”Journal of Applied Econometrics, 25, 595–620. [23]

Giacomini, R., and White, H. (2006), “Tests of Conditional Predictive Ability,”

Econometrica, 74, 1545–1578. [22,23]

(4)

24 Journal of Business & Economic Statistics, January 2015

Kiefer, N. M., and Vogelsang, T. J. (2005), “A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests,”Econometric Theory, 21, 1130–1164. [23]

Li, J., and Patton, A. J. (2013), “Asymptotic Inference about Predictive Accuracy using High Frequency Data,” working paper, Duke University. [23] McCracken, M. W. (2007), “Asymptotics for Out-of-Sample Tests of Granger

Causality,”Journal of Econometrics, 140, 719–752. [23]

Newey, W. K., and West, K. D. (1987), “A Simple, Positive Semidefinite, Het-eroskedasticity and Autocorrelation Consistent Covariance Matrix,” Econo-metrica, 55, 703–708. [23]

Patton, A. J. (2011), “Volatility Forecast Comparison using Im-perfect Volatility Proxies,” Journal of Econometrics, 160, 246–256. [22]

Rossi, B. (2005), “Testing Long-Horizon Predictive Ability with High Persis-tence, and the Meese-Rogoff Puzzle,”International Economic Review, 46, 61–92. [22]

West, K. D. (1996), “Asymptotic Inference about Predictive Ability,” Econo-metrica, 64, 1067–1084. [22,23]

White, H. (2000), “A Reality Check for Data Snooping,”Econometrica, 68, 1097–1126. [22,23]

Rejoinder

Francis X. D

IEBOLD

Department of Economics, University of Pennsylvania, Philadelphia, PA 19104 ([email protected])

I am grateful to all discussants for their fine insights. Obviously, it would be inappropriate and undesirable to attempt to address all comments here. Instead I will go to the opposite extreme, simply offering brief remarks on those that resonated most with me. In doing so I will implicitly extract what I view as an emerging theme, namely that important foundations for split-sample comparisons are emerging, as old hunches are for-mally verified (e.g., robustness to data mining) and compelling new motivation is provided (e.g., use with vintage data).1

Patton emphasizes that a key DM vs. WCM distinction (per-hapsthekey distinction) is not so much the limiting distribution theory and the conditions needed to obtain it, but rather the hypothesis being tested. In particular, DM comparisons are at estimated parameter values, whereas WCM comparisons are at pseudo-true parameter values. This is an under-appreciated yet very important point, providing a different and compelling per-spective on the appeal of the DM approach. I am grateful to Patton for raising and emphasizing it.

Inoue and by Killian are unified in their implicit emphasis on negative aspects of split-sample DM-type tests for model selection, insofar as a large price may be paid in terms of power loss, with nothing gained in terms of finite-sample robustness to data mining. Inoue and Kilian have been sounding that alarm for many years, and my article clearly reflects their influence on me. However, my pendulum has recently swung back a bit. Hence, my Section 6, which gives reasons for split-sample analysis. My

1I use “split sample” here and throughout as a catch-all term that includes

pseudo-out-of-sample comparisons based not only on truly split samples, but also on expanding or rolling samples.

reason 6.5, for example, counters the view that nothing is gained in terms of finite-sample robustness to data mining; recent work suggests that there is a gain, even if split-sample methods don’t provide complete insurance. And big data (e.g., high-frequency data), which I neglected to mention in the article, counter the view that a large price is paid in terms of power loss, as half of a huge sample is still huge.

Other discussants support my view that the pendulum is swinging back. Hansen and Timmermann, for example, em-phasize emerging work that promotes split-sample analysis. In particular, they clearly show how and why split-sample meth-ods can provide superior robustness to data mining, strongly supporting my split-sample reason 6.5 (which of course derives largely from their work). I applaud their insights and look for-ward to more.

Wright emphasizes a completely different but equally com-pelling reason for split-sample analysis: real-time comparisons using vintage data. He argues that in real-time environments with vintage data, split-sample methods are intrinsically rele-vant, and perhaps even uniquely relevant. But should not all compelling comparisons then use real-time vintage data? And if so, should notallcompelling comparisons then be split sam-ple? Put differently, if we overcome our laziness and henceforth move from “final-revised comparisons” to best-practice “vin-tage comparisons,” shouldn’t we make a corresponding move to split-sample comparisons? Time will tell.

© 2015American Statistical Association Journal of Business & Economic Statistics

January 2015, Vol. 33, No. 1 DOI:10.1080/07350015.2014.983237

Referensi

Dokumen terkait

Untuk Memperoleh Gelar Magister Teknik Dalam Program Studi Magister Teknik Elektro Pada Fakultas Teknik Universitas Sumatera

pada katrol tetap gaya yang diperlukan untuk mengangkat beban sama dengan gaya berat. beban

Menjalin Relasi dengan Sponsor/Instansi/ Kampus lain Mempermudah koordinasi komunikasi publik eksternal kampus Instansi pemerintah, pengusaha, Perbankan, kampus sahabat,

Berdasarkan Berita Acara Evaluasi/Penilaian Penawaran Administrasi dan Teknis Nomor 012/POKJA- DISHUBKOMINTEL-ACEH/V/2011 dan Berita Acara Perhitungan Kombinasi Usulan

PEMBERITAHUAN AKHIR MASA SANGGAH HASIL PELELANGAN Nomor : 027/38-04/P.SANGGAH/ULP-POKJA.BJL/SETDA/2013 Nama Pekerjaan : Pengadaan Belanja Barang Pakai Habis Medis RSU (Lelang

MODEL SISTEM INFORMASI KEUANGAN Sistem Informasi Akuntansi Subsistem Audit Internal Subsistem Intelegen Keuangan Data base Pemakai Sumber Eksternal Sumber Internal Subsistem

The minimum expected count is 19,36... The minimum expected count

Sedangkan untuk bagian dunia lainnya, konsumsi fatty acid adalah sebesar 1 juta ton pada tahun 2011 dan diperkirakan meningkat dengan rata-rata pertumbuhan sebesar 7