A Primer on Probabilistic Planning and Forecasting
Deterministic vs Probabilisitc - free to re-use with credit and link to this article

A Primer on Probabilistic Planning and Forecasting

I have been blogging and advocating for the past 15 years on probabilistic approaches to planning and forecasting, and am happy to see in the last few years it has finally started gaining traction and attention. But many questions and misperceptions still abound, about what it is and what is required. I could fill a large textbook to explain it all in detail (and I am in the process of doing just that), but here I will touch on a few key aspects.

What Makes it Probabilistic?

The critical criterion that makes a plan or forecast "probabilistic" is that the internal mathematics work on probability distributions instead of exact numbers for any value representing something uncertain. Generally, this means anything in the future. This could include quantities, lead times, production rates, yields, and so forth. The opposite of probabilistic - "deterministic" - uses exact numbers to approximate uncertain amounts, often some historical average.

A symptom of a probabilistic plan or forecast is that its results are generally also expressed as probability distributions. Where a deterministic plan or forecast is generally expressed as a set of time series of exact numbers, a probabilistic plan or forecast is expressed as a set of time series of probability distributions. It is important to realize even results of deterministic plans can be expressed as distributions, but that does not imply that they are probabilistic. For example, fitted error of statistical forecasts can be expressed as a distribution, assuming it has a normal distribution centered around the forecasted number. The output can then be presented as a "range forecast". But this should not be confused with a probabilistic forecast. A range forecast that is statistically determined will generally have a large error due to the naive assumption of the distribution being independent, identically distributed (i.i.d.) and Gaussian (aka having a normal distribution), which is rarely true in the real world.

As an example, consider repeatedly throwing two dice. An unbiased statistical forecast will state that you will throw 7 every time, since that is the expected average. A probabilistic forecast will be expressed as various probabilities of throwing any potential outcome:

No alt text provided for this image

Figure 1: a statistical forecast of throwing 2 dice (left) and its probabilistic equivalent (right)

The "error" on the statistical side is not really an error of the forecast at all. It is the inability to distinguish between true error and natural variability.

If we show this as time series, both would be level (stationary) since the outcome is identically distributed each throw. More typical for supply chain data they would look similar to this:

No alt text provided for this image

Figure 2: a statistical forecast (left) and a probabilistic one (right) for weekly data. Time is horizontal, forecast quantity vertical. Lighter shades of blue indicate lower probability of occurring (a visual simplification for easier interpretation). The red dashed lines are typical confidence ranges or range forecast.

The Value of Probabilistic Methods

The root of the value of the probabilistic approach is that it can properly distinguish between error and natural variability, and between signal and noise, which is impossible in the deterministic perspective. This has three main consequences:

  1. Risk and opportunity are impossible to determine accurately from deterministic plans and forecasts.
  2. It is impossible to properly judge deterministically how good or bad a plan or forecast is.
  3. It is impossible to determine with any degree of accuracy where to focus improvement efforts based on deterministic plans or forecasts.

A symptom of these limitations are the various metrics used. If you have a MAPE of 60% , is that good or bad? And if you believe it is bad, how much better could you make it? Honestly, that is impossible to say.

Probabilistic approaches provide rich information to identify risks and opportunities at all levels of detail, allowing informed business decisions to be made. They also allow perfect delineation of the things you can control and improve versus the things you cannot. It is possible to create metrics (probabilistic metrics!) that perfectly quantify signal versus noise, error versus variability, precision versus accuracy.

Precision versus Accuracy

This brings us to the topic of precision and accuracy. These terms have been used in the forecasting domain for over half a century, and often interchangeably. Frankly, they have been used wrong. They have different meaning in the forecasting domain than every other domain. And this is due to the deterministic perspective not being able to distinguish between the two. So, arbitrary convolutions of the two were historically assigned names that in other domains have pure and clear meanings. The probabilistic perspective allows reclaiming these terms and using them in the generally correct way:

No alt text provided for this image

Figure 3: precision versus accuracy

In the pure perspective, precision and accuracy are orthogonal: they say something completely different about the same thing.

  • Accuracy - how close to target is the plan/forecast on average
  • Precision - how close are the plan/forecast values to each other

Accuracy can be expressed on a known scale (such as 0% to 100%) where each value is really possible, but it cannot be determined at time of forecasting. Precision can be determined at time of forecasting or planning because it is independent of the actual values, but it is on an unknown scale. Deterministic metrics can never be determined before actual values are known, and most are on an unknown scale. So-called in-sample approximations tend to severely underestimate true values. The worst of both downsides.

For more details see for example "How to Distinguish Between Accuracy and Precision in Forecasting" or "Three Types of Precision and Their Usefulness in Forecasting".

When we apply these concepts to the example in figure 1, it should be clear that statistical forecasts are precise but inaccurate. They are expressed as exact numbers, which are generally wrong. On the other hand, probabilistic forecasts can be perfectly accurate but are imprecise. They are expressed as vague ranges, but these can be true representations of reality. They are polar opposite approaches! You cannot tweak one to get the other. They are fundamentally different.

How is this Different from Scenarios?

Two common misconceptions are that probabilistic plans are like range plans or like different scenarios. There is some overlap for sure. A probabilistic plan expresses the full range of probabilities for a known or determined set of assumptions and actions. For example, what is the range of possible demand values given a current promotion plan? Under the assumptions, it can tell you things like worst case, best case, and most likely. This is different from scenarios for different input assumptions, such as what if we decreased or increased our promotional budget. In the probabilistic perspective each of those will have its own ranges from worst to best case:

No alt text provided for this image

Figure 4: Deterministic (Point, Range, top) versus Probabilistic (bottom) perspectives of forecasts (left), plans (middle), and scenarios (right).

This figure illustrates the two different types of scenarios. The range of possible outcomes given current assumptions can be provided using range or probabilistic outputs. The what-if possible outcomes given different sets of assumptions can be provided using multiple point plans or probabilistic plans. The combination of the two: many possible ranges of outcomes for multiple possible what-if sets of assumptions can only be provided in a controlled way by probabilistic approaches (as opposed to many rough approximations via brute force with large computing power).

A Word on Closed-Form and Simulation

None of the above makes any assumption on the techniques used to generate probabilistic plans. The choice is one between easy (deterministic) and accurate (probabilistic). But getting accurate can be very difficult. Three most common ways to generate probabilistic plans and forecasts are:

  1. Use empirical distributions (sample historical data) and simulate future possibilities.
  2. Use assumed distribution fitted to historical data and perform closed form calculations.
  3. Use assumed distribution fitted to historical data and simulate future possibilities.

Option 1 works really well when there is large amounts of data for each and every time series to be planned or forecasted, but requires huge amounts of computing resources as data increases in size. When data is sparse, for example for intermittent demand items, the empirical distributions deteriorate and results become unreliable. This option is often referred to as "stochastic".

Option 2 works really well when the data can be accurately classified to certain parametrized distributions. It requires no more data than deterministic equivalent methods and runs on same computing resources. Downside is that it is extremely difficult to do, since it requires probabilistic arithmetic, which has no known general purpose examples, but for specific families of distributions it can be done. The Gaussian (or normal) distribution is one case where the math is pretty simple and hence why it is used so often. However, because it is such a poor fit to the real world, it is never used by probabilistic systems. The purpose is accuracy, remember, not ease...

Option 3 is a blend of the other two. Using an assumed distribution to avoid the problem with sparse data, and using that in simulations to avoid the difficult math, at the expense of requiring large amounts of computing resources for meaningfully sized problems to solve.

Other variations are also encountered. Most are blends or enrichments of the above options.

Myths about Probabilistic Methods

There are a few misconceptions that are regularly raised when talking about probabilistic methods that are beyond the scope of this primer. I will likely write additional articles on each of these in the near future. For now, I list them here:

  • Myth 1: probabilistic methods need large amounts of data. False: most data exists in every ERP system and can be supplemented in a similar fashion to its deterministic counterparts (such as promotional data). It can include much more data if desired and available to further isolate signal from noise, but that is optional.
  • Myth 2: probabilistic methods need huge amounts of computing power. Only true if using the brute force approach of simulation for large data sets. False for closed-form methods, which uses no more than deterministic equivalents, and often less.
  • Myth 3: Probabilistic plans and forecasts are difficult to interpret and use. False, they tend to provide richer information, allowing better-informed business decisions. Where deterministic plans provide exact average quantities with an understood but unquantified uncertainty, probabilistic methods present the uncertainty accurately. Discussions change from politics to ones where the balance between cost and risk is decided.
  • Myth 4: Probabilistic plans and forecasts cannot be adjusted by humans. False. In fact, adjustments can be made exactly as desired, but the uncertainty skews when historical adjustments have been biased. This allows perfect alignment to biased sales forecasts or budgets without exploding supply chain costs to meet those.
  • Myth 5: Artificial Intelligence (AI) and Machine Learning (ML) are probabilistic. Mostly false. There are some AI methods that could be considered probabilistic, although most of those naively assume a normal distribution applies. The vast majority are purely deterministic. That said, AI/ML can be used to complement probabilistic methods where those are weak, for example when there is no or little historical data.

Conclusion

Probabilistic methods are the future. They are the correct foundational building block for domains that contain uncertainty. This includes all types of forecasting and all types of planning processes. The traditional deterministic methods assume incorrectly that uncertain values can be safely approximated by some average single number. Forecasts that are always wrong, plans that are infeasible by the time they are published because things changed, and perpetual expediting and fire-fighting are the direct consequences of this one assumption. These methods are fundamentally flawed and cannot be salvaged. They are precise when accuracy is what is needed.

The fundamentally correct way, probabilistic methods, however, are difficult to develop. It will require much greater sophistication from software and software developers in order to make their solution much less complex than traditional systems for the planners and forecasters.

It will also require some epiphanies from all involved. Not least that accuracy and precision have been bastardized in the forecasting domain, and need to be reinvented to match all other domains. But no less that all deterministic metrics are inadequate to the needs because they fail to capture this distinction cleanly. Many still have value in the probabilistic worldview, but not how they are used today. More on that in future articles.

If you are interested in probabilistic planning and forecasting please consider joining the "Probabilistic Supply Chain Planning" group here on LinkedIn.

All content is original by Stefan de Kok. Feel free to re-use in your own materials as long as credit is provided and/or a link to this article is included.

Find all my articles by category here. Also listing outstanding articles by other authors.

José Villamil

Supply Chain Management Expert | Team Leader

3y

Very interesting; I am start searching this topic for my thesis, and I'm not decided yet if I should go in the direction of probabilistic planning or big data (if I can use such thing for material planning with my limited resources) for modeling the behavior of demands and stocks. In any case its clear in my opinion that the uncertainty of today and the velocity of changes requires some fuzzy logic to predicts the stocks and flows along any global supply chain.

Jean-Baptiste Clouard

CEO and co-founder chez Flowlity

3y

Great article!

Sachin Shedge

Head of Supply Chain - Asia Pacific at Zespri International

3y

Hi Stefan - Indeed a nicely articulated and thought provoking article. Quick question. What about forecasting of high value items like gold, platinum or high end medical devices? Which approach will be more effective?

To view or add a comment, sign in

Insights from the community

Others also viewed

Explore topics