what to compare
how ro compare
graphical
diagnostic tools;statistics
statistics
Data | Methods |
---|---|
all data |
\(\chi^2\), GAIC |
test data |
LS, CRPS, MSE etc. |
data partitioning |
|
K-fold | LS, CRPS |
bootstrap | LS, CRPS |
no partition
evaluation is done in the training dataset
likelihood ratio
test for nested models \[LR= GD_1/ GD_0 \sim \chi^2(df0- df_1) \]partition
: evaluation is done on new
data
LS
) \(\equiv\) Prediction Global Deviance (PGD
)CRPS
)MAPE
)models | df | AIC | \(\chi^2\) | BIC |
---|---|---|---|---|
mlinear | 24 | 22808.89 | 22855.45 | 22953.69 |
madditive | 29.24 | 22814.33 | 22871.07 | 22990.77 |
mfneural | 160 | 22930.36 | 23240.76 | 23895.69 |
mregtree | 30 | 38754.64 | 38812.84 | 38935.64 |
mcf | 404 | 38754.64 | 24481.28 | 26134.99 |
models | LS | CRPS | |
---|---|---|---|
mlinear | -6.2302 | 73.5738 | |
madditive | -6.2288 | 73.8201 | |
mfneural | -6.5126 | 79.9394 | |
mregtree | -6.2930 | 78.6704 | |
mcf | -6.2966 | 79.0644 |
the GAIC is well established and tested (the df of freedom need to be known)
the linear and additive model are good when there are not many explanatory variables (but somehow interactions have to be considered)
more work has to be done to standardised all ML techniques so their partitioning of data are comparable to the conventional additive models
The Books
www.gamlss.com