Results and discussion Directory UMM :Data Elmu:jurnal:E:Economics of Education Review:Vol18.Issue4.Oct1999:

409 B.D. Baker, C.E. Richards Economics of Education Review 18 1999 405–415 were applied. 15 For Group Method of Data Handling GMDH, default parameters were also applied. 16 For both the recurrent backpropagation models and Generalized Regression neural networks, the training sets consisted of data from 1960 to 1980 and test sets consisted of data from 1981 through 1990. 17 Alternate models were trained with log ln transformed data. Pre- dictors and lags were the same as those specified for the linear regression equation. Only the main equation for the prediction of CUREXP was modeled with the neural networks. Forecasts were generated for the period from 1991 through 1995 by providing the trained neural net- works with the same set of forecast predictors univariate results used for the linear regression model including regression predictions of SGRNT from the secondary multiple linear regression equation.

4. Results and discussion

4.1. Univariate models All models selected for univariate forecasting were either 1 Linear Holt Exponential Smoothing 18 PCI, PERTAX or 2 Damped Trend Exponential Smoothing ADAPOP, BUSTAX, RCPIANN models. Parameters for the models and measures of fit are displayed in Appendix B. 4.2. Multivariate model The estimated AR1 multivariate regression model for forecasting SGRNT is: ln SGRNT 5 0.13 0.90 15 Including linear scaling 0,1 of the input layer and model selection based on test set error, with the stop criterion set to 20 generations without reduction of error. One alteration made was increasing the size of the gene pool to 300 from the default of 20. Incrementally increased gene pool size, while requiring longer training times, produced more accurate results. 16 Including linear scaling of the input , 2 1, 1 1 and output layers [ 2 1, 1 1], “Smart” model selection and optimiz- ation, applying FCPSE as the selection criteria, and the option to construct a model of “high” non-linearity, indicating the option to include second- and third-order terms on all para- meters and two- and three-way interactions. 17 An arbitrary specification of 33, in excess of the rec- ommended 20 WSG, 1995, p. 101 test set that yielded equi- valent results. 18 Which for prediction or forecasting purposes can be expressed as: Z ˆ n l 5 a[Z n 1 1 2 aZ n 2 1 1 1 2 a 2 Z n 2 2 1 … ] 5 S n , where the predicted future observation Z ˆ n l is based on previous observations Z n − t multiplied by a decaying smoothing constant a, 1 2 a 2 … and S n represents the smoothed statistic Liu Hudak, 1994, p. 9.3. 1 0.67 ln BUSTAX{1} 0.04 1 0.30 ln PERTAX{1} 0.12 3 1 0.31 ln ADAPOP 0.01 2 0.03 lnRCPIANNRCPIANN{1} 0.08 with Rho 5 0.19 0.37, R 2 5 0.992, p-values in parenth- eses. Forecasts of SGRNT were generated from this model using the univariate forecasts of predictors BUSTAX, PERTAX, ADAPOP, RCPIANN Appendix C. The estimated AR1 multivariate regression model for forecasting CUREXP is: ln CUREXP 5 2 1.30 0.29 1 0.55 ln PCI 0.02 1 0.63 ln SGRNT 0.00 4 2 0.32 ln ADAPOP 0.02 with Rho 5 0.41 0.04, R 2 5 0.997, p-values in parenth- eses. Forecasts of CUREXP were generated from this model using the univariate forecasts of PCI and ADAPOP, and multivariate forecast of SGRNT Appendix C. 4.3. Neural models Table 1 displays the training sample fit and pro- duction forecast fit results for each of the neural net- work models, including both fitness measures and each network’s respective measure of input contribution to prediction. For the recurrent backpropagation models, each obtained exceptionally strong fit to the training set R 2 5 0.993 to 0.994ln and produced reasonable fit to the production set R 2 5 0.756 to 0.757ln. The aggre- gate contribution factors 19 implicate state-level support SGRNT as the primary influence on spending. Both GRNN models were slightly less fit R 2 5 0.910 to 0.956ln to the training set than recurrent backpro- pagation models and both produced looser fit to the pro- duction set 0.643 to 0.688ln. Again, aggregate weighting factors, in this case smoothing parameters, suggest relatively high importance of state-level support SGRNT for predicting spending CUREXP. GMDH produces comparable fitness to both training and production sets to recurrent backpropagation. Note, 19 Like regression coefficients, these factors represent a meas- ure of the responsiveness of the outcome measure to changes in inputs, yet, as aggregate contribution factors they express no direct mathematicalfunctional relationship between inputs and outputs. Rather they indicate a rough measure of relative impor- tance of inputs derived from the weighting matrix. For a more detailed discussion, see WSG 1995, p. 191. 410 B.D. Baker, C.E. Richards Economics of Education Review 18 1999 405–415 Table 1 Training and forecast fitness of neural models Training R 2 Production R 2 PCI ADAPOP SGRNT Backpropagation 0.993 0.756 0.289 a 0.078 0.525 Backpropagation ln 0.994 0.757 0.310 0.062 0.517 GRNN 0.910 0.643 2.318 b 0.024 2.694 GRNN ln 0.956 0.688 2.141 0.035 2.765 GMDH 0.983 0.793 0.99 c – – GMDH ln 0.997 0.740 – 0.21 d 0.84 e PCI: per capita income; ADAPOP: average daily attendance as a percent of the population; SGRNT: state-level support to public education per pupil in ADA a Aggregate weighting or “Contribution” factor see WSG, 1995, p. 191. b Aggregate smoothing factor a see WSG, 1995, pp. 198–205. c Linear coefficient b, based on rescaled PCI ˆ PCI 5 2 3 PCI 2 7504.3913932.61 2 1. d Linear coefficient b, based on rescaled SGRNT ˆ SGRNT 5 2 3 ln SGRNT 2 4.791.28 2 1. e Linear coefficient b, based on rescaled ADAPOP ˆ ADAPOP 5 2 3 ln ADAPOP 1 1.980.48 2 1. however, that the GMDH model that produces greater fit to the training data 0.997ln compared to 0.983 does not perform as well on the production data 0.740ln compared to 0.793. In addition, while measures of pre- dictor importance were fairly consistent between back- propagation and GRNN models, GMDH models took varied forms, with the first not log transformed includ- ing only per capita income and the second log transformed including only state support and average daily attendance. Interestingly, though GMDH learning was set for the possibility of including both second- and third-order terms as well as two- and three-way interac- tion terms, resultant best predicting models included only first-order linear parameters. 4.4. Model forecasts and accuracy assessment Table 2 displays the AR1 and neural forecasts of edu- cational spending alongside actual spending values 1982–1984 constant dollars for the period 1991–1995. Note that beyond 1993 all models provide overestimates Table 2 Comparison of forecast accuracy Group method of data Recurrent backpropagation Generalized regression handling Year Actual US AR1 US A US B a US A US B a US A US B a US 1991 3923 3856 3869 3853 3934 3917 3870 3870 1992 3920 3918 3901 3886 3949 3947 3906 3906 1993 3916 3973 3934 3925 3994 3970 3942 3942 1994 3942 4027 3970 3968 4026 4017 3977 3978 1995 3979 4081 4004 4010 4029 4061 4013 4013 Prices reported in constant school-year 1982–84 dollar using reported consumer price index CPI as the deflator. a Indicates model trained with log ln transformed series. of educational spending, the most significant overesti- mates coming from the AR1 model. Table 3 displays the comparisons of forecast accuracy for the AR1 multiple regression model and the three neu- ral network models. While the NCES model performed quite well for the entire period MAPE 5 1.60, each of the neural models outperformed the NCES model on average for the forecast period. Period-to-period predic- tions, however, vary quite significantly, for example, with the AR1 model outperforming the others for the second 1992 time period. Overall, recurrent backpro- pagation and GMDH models performed particularly well, providing an approximate 2 3 increase in forecast accuracy over the NCES model. Despite less accurate overall predictions of the level of spending, GRNN models yielded the only results sug- gesting non-linear sensitivity, revealing more accurate changing rates of growth in spending from 1991 to 1993, given the relatively constant univariate extrapolations of predictors. During this period, actual spending declined 2 0.1, AR1 forecasts increased 1.5, and on aver- 411 B.D. Baker, C.E. Richards Economics of Education Review 18 1999 405–415 Table 3 Comparison of forecast accuracy Recurrent backpropagation Generalized regression Group method of data handling Year AR1 A B a A B a A B a 1991 1.71 1.38 1.78 0.27 0.18 1.37 1.36 1992 0.07 0.50 0.88 0.74 0.67 0.38 0.36 1993 1.47 0.47 0.23 2.00 1.38 0.66 0.68 1994 2.17 0.72 0.68 2.14 1.92 0.91 0.92 1995 2.57 0.64 0.80 1.26 2.06 0.87 0.87 MAPE b 1.60 0.74 0.88 1.28 1.24 0.84 0.84 a Indicates model trained with log ln transformed series. b Among others, Chase 1995 explains the interpretation and use of Mean Absolute Percentage Error MAPE in assessing forecast accuracy. age backpropagation and GMDH forecasts increased 0.9. GRNN forecasts increased more slowly from 0.7 to 0.8, with a low of 0.4 for 1991–1992 GRNN and 0.6 for 1992–1993 ln GRNN. That GRNNs revealed this pattern is consistent with Caudill’s Caudill, 1995, pp. 47–48 suggestion that GRNNs are particularly sensitive to non-linear relationships in sparse andor noisy data.

5. Conclusions and recommendations