Week 7:
Prediction Intervals and Advanced Forecasting Options

Agenda

Review of forecasting

Understanding prediction intervals

Time-series cross-validation

Project 1 reminder and details

Review: Forecasting and Decomposition

Let’s recall the process of breaking down time series data.

  1. We separated (through time-series decomposition) a times series into:
  • Trend: long-term direction
  • Seasonality: repeating patterns
  • Residual: unexplained variation
  1. We used forecasting to predict trend and seasonality into the future.

  2. We modeled residuals separately using an ARIMA model.

  3. Final forecast = Trend + Seasonality + Residual Model

Intro to Prediction Intervals

What are prediction intervals?

The interval within which we expect
values of our \(y\) variable to lie at
a future point in time.

Prediction intervals: An example

Last week, we forecasted carrot prices from August 2024 to January 2028.

The predicted price for:

  • April 2025 is $261.5564
  • January 2028 is $255.9185

But how confident are we? Instead of a single value, we might say:

  • We are 95% confident that the April 2025 price will be between $142.30237 and $380.8104.
  • We are 95% confident that the January 2028 price will be between $28.73887 and $483.0982.

These are our prediction intervals.

  • Prediction intervals increase as we forecast further into the future.
  • Prediction intervals capture both model uncertainty and random fluctuations.
  • Wider intervals = more uncertainty in the forecast.

Before we introduce prediction intervals, let’s first understand confidence intervals (CIs).

What are confidence intervals?

  • A CI is a range of values that express uncertainty in an estimate.
  • In forecasting, a CI tells us the range where the true trend or seasonal effect is likely to fall.
  • If we estimate a trend in carrot prices, a CI might say:

“We are 95% confident that the true price trend is between $0.05 and $0.10 per month.”

How are confidence intervals calculated?

CIs are calculated using:

  • The estimate itself (e.g., the trend or seasonal component),
  • The standard error (how much the estimate varies),
  • The sample size (more data = narrower intervals),
  • A multiplier (critical value) (based on the confidence level).

How are prediction intervals different?

  • Confidence intervals apply to estimates like trends or seasonal effects.
  • Prediction intervals apply to individual future values of a time series.

Example:

  • A confidence interval might tell us how uncertain we are about the trend in carrot prices.
  • A prediction interval tells us how uncertain we are about the actual price of carrots next year.

Prediction intervals are always wider than confidence intervals because they capture both:

  • Uncertainty in the estimate (trend, seasonality, etc.).
  • Uncertainty in future random fluctuations (residuals).

Interpreting prediction intervals

The most frequently used prediction (and confidence) interval is 95%.

The intuitive interpretation: We are 95% confident that our prediction interval contains the value of our outcome variable that will be realized.

In Tableau (and most other software) prediction intervals are shown with shaded regions around the prediction.

Review Question 1

Which of the following best defines confidence intervals?

A. A range of values that is likely to contain an unknown population parameter with a specified level of confidence.

B. A range of values that represents the mean of a sample.

C. A range of values used to estimate a future value of a variable with a specified level of confidence.

D. A range of values that is used to test a hypothesis about a population parameter.

Review Question 2

Which of the following best defines prediction intervals?

A. A range of values that is likely to contain an unknown population parameter with a specified level of confidence.

B. A range of values that represents the mean of a sample.

C. A range of values used to estimate a future value of a variable with a specified level of confidence.

D. A range of values that is used to test a hypothesis about a population parameter.

Prediction intervals in R

Last week, we generated forecasts using forecast() in R.

Let’s look at the prediction interval for the carrot forecast:

The shaded region shows the prediction interval range.

How are prediction intervals calculated?

How are prediction intervals computed (95% level)

The formula for a 95% prediction interval for the \(h\)-step forecast is:

\[\text{Prediction Interval for }{y}_{T+h|T} = \widehat{y}_{T+h|T} \pm 1.96 \cdot \widehat{\sigma}_h\]

where

\(\widehat{y}_{T+h|T}\) is our prediction of \(y\) at time \(T\), \(h\) periods into the future

\(h\) represents the forecast horizon, or the specific number of intervals (e.g., days, months, years) from the last observed data point

1.96 is the multiplier for a 95% prediction interval assuming that our sampling distribution of future observations is normal

\(\widehat{\sigma}_h\) is an estimate of the standard deviation for the \(h\)-step forecast

Understanding the multiplier in prediction intervals

We adjust the interval width based on how confident we want to be:

Confidence Level Multiplier
80% 1.28
85% 1.44
90% 1.64
95% 1.96
99% 2.58
  • Higher confidence \(\rightarrow\) wider prediction intervals
  • Lower confidence \(\rightarrow\) narrower intervals

How are prediction intervals computed (general form)

The equation for a prediction interval can be written more generally as: \[ \widehat{y}_{T+h|T} \pm c\cdot \widehat{\sigma}_h \]

where the multiplier \(c\) depends on the prediction interval you desire.

Note that these all assume a normal distribution of future observations.

Why does standard deviation matter for prediction intervals?

  • The standard deviation of the forecast errors measures how much our predicted values tend to vary from actual observations.

  • The further we forecast into the future, the greater this uncertainty becomes.

  • We adjust our prediction intervals based on the standard deviation of the forecast distribution.

Let’s look at how this is done mathematically.

Prediction interval formula: Standard deviation

When forecasting 1 step ahead (\(h=1\)), we will estimate the standard deviation of the forecast distribution according to:

\[ \widehat{\sigma} = \sqrt{\frac{1}{T-K}\sum^T_{t=1}e^2_t} \]

where

\(T\) is the number of data points in the time series

\(K\) is the number of parameters estimated in the forecasting method

\(e_t\) is the residual at time \(t\)

A note on residuals

The residual \(e_t\) in the equation of the standard error is not the same as the residual \(R_t\) from the time series decomposition.

Residuals (\(R_t\)) in Time Series Decomposition

  • Represent the portion of the data that is not explained by the trend and seasonal components.
  • Captures random fluctuations or noise in the observed time series.

Residuals (\(e_t\)) in Forecasting & Prediction Intervals

  • Represent the errors between predicted and actual values when making forecasts.
  • Measure how much a model’s predictions deviate from reality.
  • Used to estimate the uncertainty in future forecasts.

Prediction interval formula: Standard deviation

When forecasting \(h\) steps ahead (\(h\geq 1\)), we can estimate the standard deviation of the forecast distribution according to:

\[ \widehat{\sigma}_h = \widehat{\sigma}\sqrt{h} \]

This formula highlights how uncertainty grows over time.

Note that there are many other methods for estimating the standard deviation of the forecasting distribution. This formula is for the Naïve method, and is among the simplest.1

Prediction intervals: Application

In lab this week, we will show you how to estimate prediction intervals in R and plot them in Tableau.

Validating Predictions

How do we validate
our forecast model?

The objective:
We want to show our audience
that they should trust our predictions.

The solution:
Time series cross-validation
evaluates forecasting accuracy.

The process

Train-Test Splits

  • Use historical data to train a forecast model.
  • Compare predictions to held-out test data.

Measure Accuracy

  • Compute Mean Absolute Error (MAE), Root Mean Square Error (RMSE), etc.

Train-test splits

Construct a series of “training” and “test” sets of data.

Use your “training” set to construct a forecast (following the same methods in your main forecast).

Evaluate the accuracy of your forecast using your “test” set.

Example: Train-test splits

Training data = blue

Test data = orange

Analogy: Predicting the Weather with a Growing Diary

Imagine you’re keeping a diary of the daily high temperature in your town.

  • At first, you have only a few days recorded (blue dots), so you try to predict tomorrow’s temperature (orange dot) based on what you’ve seen so far.
  • Each day (each additional row), you update your diary by adding yesterday’s temperature and then make a new prediction for today (the previous row’s orange dot).
  • As time passes, your diary grows, and you always use all past records to make the best possible guess for the next day.

\(\rightarrow\) Each day, you train a model on all previous days and test your prediction on the next.

\(\rightarrow\) Over time, your predictions (hopefully) get better as you learn from more past observations!

That’s exactly how rolling-origin cross-validation works in time series forecasting.

Measure accuracy

How do we measure how well our model predicts future values?

Metric Formula Interpretation
Mean Forecast Error (MFE) \(\sum_{t=1}^{H} \frac{\hat{y}_t - y_t}{H}\) Average difference between predictions & actual values.
Mean Absolute Error (MAE) \(\sum_{t=1}^{H} \frac{| \hat{y}_t - y_t | }{H}\) Measures average absolute error size.
Root Mean Squared Error (RMSE) \(\sqrt{\sum_{t=1}^{H} \frac{ (\hat{y}_t - y_t )^2 }{H}}\) Penalizes larger errors more heavily.
Mean Absolute Percentage Error (MAPE) \(\sum_{t=1}^{H} \frac{ (\hat{y}_t - y_t )/y_t }{H}\) Expresses error as a percentage of actual values.

Choose the forecasting model that minimizes prediction error according to these measures.

This is a useful resource for these evaluation methods: https://medium.com/analytics-vidhya/basics-of-forecast-accuracy-db704b0b001b

REMINDER: Project 1

Groups of 2

Choose an ag biz or enre management question to answer with time series data

Collect time series data

Analyze trends

Generate a forecast

\(\rightarrow\) Generate prediction intervals

Present results in a recorded video

EXAMPLE: The Question

I am currently growing almonds and walnuts, and in the coming season I will plant a new orchard.

  • Almond trees need to mature for 3-4 years before they are ready for harvest.

  • Walnut trees need to mature for 4-7 years before they are ready for harvest.

  • Both trees will continue producing for roughly 25 years.

How will I decide which trees to plant?

EXAMPLE: The Data and Analysis

DATA: Time series data on almond and walnut prices.

ANALYSIS: Forecast prices for the next 10 years (really 25 would better answer your Q, but is probably too long of a prediction period)

ANSWER: I will plant walnuts because over a 10 year horizon they would yield higher revenues than almond trees, even though I would have to wait longer before selling them.

Presentation Overview

Your presentation videos should be roughly 8 minutes. (Minimum: 6 min, Maximum 9 min)

You should begin your presentation with the following:

  • Who are you? (what business, organization, etc. wants an answer to your question?)

  • Who is your audience? (business executive team, policy makers, consumers?)

You should then:

  • Give us any relevant background info that is necessary to understanding your question

  • State your question

Presentation Content

Data:

  • Describe where your data is from

  • Provide summary statistics and visualizations

Analysis:

  • What forecasting model did you use? Why?

  • Show your forecast and prediction intervals

Discussion/Conclusion:

  • How did you use your forecast to answer your question?

  • What are the key take-aways for your intended audience?

  • What are limitations to your analysis?

Final Project Submission

The video presentation is part 1 of the project.

A report summarizing the presentation content is part 2 of the project. Your report should summarize the same information as your presentation, and can include additional information and visualizations if they are helpful for answering your question.

Each group member will be responsible for uploading the report onto their Google site. You can both have an identical report, but both must post it to your personal webpage.

References

Hyndman, R.J., & Athanasopoulos, G. (2021) Forecasting: principles and practice, 3rd edition, OTexts: Melbourne, Australia. https://otexts.com/fpp3/. Accessed on 02-28-2023.

A useful blog post: https://mins.space/blog/2020-07-30-forecasting-through-decomposition/

Another useful blog post: https://medium.com/analytics-vidhya/basics-of-forecast-accuracy-db704b0b001b