Defining Simple Fitted Values
- For most time-series forecasts, each observation uses all previous observations
- Implying, each forecast for is based on observations
-
Thus, a forecast should be denoted as
- However, we usually just write because it's simpler
- Technically, fitted values aren't always true forecasts since they're estimated using observations, including those after (i.e. )
- For example, the fitted values of a moving average forecast will represent the average of all observations, where
- Also, the fitted values using the drift method uses a parameter estimated on all observations of , where
- On the other hand, naive forecasts don't include any parameters, so fitted values are true forecasts (but are less accurate usually)
Defining Residuals for Time-Series Models
- Residuals refer to the amount we're off by when calculating a prediction on a data value
- Specifically, a residual represents
- Thus, we can use residuals to evaluate the accuracy of our predictions
-
To ensure our forecasts aren't biased, we usually enforce the following:
- Residuals are uncorrelated
- Residuals have mean and constant variance
- Residuals are normally distributed
Defining Notation for Evaluation Metrics
-
Before diving into specific evaluation metrics, let's define standard notation for forecasting metrics:
- Let denote the current observation at time
- Let denote the previous observation at time
- Let denote the forecast of
- Let denote the forecast error where
- Let denote the one-step naive error where
Evaluating Forecasting Accuracy using MAPE
- The mean absolute percentage error (or MAPE) is defined as the following:
- The MAPE favors predictions that are smaller than its data value, which can be considered a drawback
- On the other hand, we may want this property depending on our problem, in which case we would want to use MAPE
- In other words, the MAPE puts a heavier penalty on forecasts that exceed the actual data values than those that are less than the actual values
- Said another way, the MAPE puts a heavier penalty on negative errors than positive errors
- Naturally, we would like to avoid this asymmetry of the MAPE
- The MASE can be used if we want a more symmetrical measure of the percentage error
Evaluating Forecasting Accuracy using MASE
- The mean absolute scaled error (or MASE) is arguably considered the best available measure of forecast accuracy
- Before we define the MASE formula, we should define a one-step naive error
- The one-step naive error refers to the error associated with guessing the previous data value as our current prediction
- The MASE is defined as the following:
- Where the scaled error term refers to the following:
- Therefore, the MASE formula can be simplifed to the following:
- We can go one step further, and simplify the MASE to the following, roughly:
- Where the scaled error term roughly refers to the following:
- Essentially, the MASE is an average of our scaled errors
-
The MASE has the following benefits:
- Working with scaled errors, since the scaled errors are independent of the scale of the data
- Symmetrical measure
- Less sensitive to outliers compared to other metrics
- Easily interpreted metric using scaled errors (compared to other metrics like RMSE)
- Less variable on small samples
-
We can interpret scaled errors based on the following criteria:
- A scaled error is less than one if our forecast is better than the average one-step naive forecast (i.e. using the previous data point)
- A scaled error is greater than one if our forecast is worse than the average one-step naive forecast (i.e. using the previous data point)
References
Previous
Next