Forecasting

Blast From The Past
This was first posted to the original incarnation of this blog on 19/02/2017. I’ve reposted here it with some formatting changes and minor edits for clarity so it is not buried in this repository’s history. This is one of the topics I hope to revisit again in more detail in the future.

This post by Neil Killick has a good summary of the dysfunctional behaviors that seem to be permanently wedded to estimation in software development. It’s a great post, and you should read it all, but I want to focus on two points that are relevant to something I’ve been mucking about with recently.

Quote
Not incorporating appropriate levels of uncertainty into an estimate (e.g. giving a precise date when a set of scope will be delivered, saying a precise set of scope will be delivered on a fixed date, not calling out risks/assumptions/issues/dependencies as part of the estimate, etc.)

and

Quote
Due to the complex nature of product development, teams should only estimate individual backlog items and forecast how much they can do in a short timespan (e.g. 2-week sprint/iteration) For batches of items beyond this short timeframe (e.g. releases), empirical forecasting using “yesterday’s weather” should be used to answer “how long” or “when” questions, not asking developers.

Monte Carlo Forecasting

An easy way to help address both of Neil’s points is to create forecasts using the monte carlo method. These forecasts are based on historical performance, can provide a range of possible outcomes with a likelihood for each and should not require any additional time from your development team.

Here is a worked example1 to explain how to make one:

  1. The first input to the forecast is historic performance data for your team, measured in some units of work per time period. This could be velocity in story points per sprint, or throughput of stories per week. For the purposes of creating a forecast, the specific units used are not important provided they are consistent across the historic data.

    To keep this example simple we’ll assume a development team with a velocity measured in story points per sprint.

    IterationVelocity
    16
    28
    35
    413
    57
  2. The next input is the number of time periods into the future that you want to forecast. These are the same time units used in the historic data described in step 1.

    We’ll assume we want to forecast our next release, which will be in 8 iterations time.

  3. The last input is the number of trials. Think of each each trial as a potential ‘future timeline’ for your team. As a gross simplification, the more trials used, the more confident you can be in the forecast2.

    Because this is an example, we’ll pick an unreasonably low number of trials. Say 5.

  4. We populate each time period in each trial by randomly sampling the historic data from step 1. This will result in a data set that looks a bit like this:

    IterationTrial 1Trial 2Trial 3Trial 4Trial 5
    11313585
    2881355
    385686
    4137755
    585685
    65513513
    7138658
    8885513
    Total7659614960
  5. We can summarize the trials in step 4 by expressing the likelihood of achieving a certain number of story points as the percentage of the total number of trials where that number of story points was achieved.

    Story Points AchievedLikelihood
    40100%
    5080%
    6060%
    7020%
    800%

    To continue with the crap sci-fi metaphor, think of this as the percentage of future timelines where our customer’s increasingly wild dreams come to pass.

Explicit Uncertainty

Probably the most common method used to create empirical forecasts is to multiply the average (mean) of a team’s historic performance by some number of time periods, and then express that as a single figure. If the team in the example above did this they would have come up with a forecast of roughly 62 story points in 8 iterations.

I can guarantee that anyone exposed to that unqualified 62 story point figure will automatically attribute a likelihood of 100%3 to it with no further comment. Unfortunately the actual likelihood is more like 50%. All kinds of important business decisions are now going to be made based on odds no better than a coin toss. Oops.

This is where the monte carlo forecast adds value. It produces a range of possible outcomes with an explicit level of uncertainty for each, better enabling ’the business’ to trade off risk and reward when making decisions.

Some Code

It’s unfashionable to do maths with your own brain, so I wrote some code to create these simple forecasts. You can find it on GitHub. This is how to use it to recreate the example above:

forecaster.exe --command percentile --samples 6 8 5 13 7 --iterations 8 --format pretty

And this is what the output will look like:

+------------+-------+
| Likelihood | Value |
+------------+-------+
|        100 |    41 |
+------------+-------+
|         90 |    52 |
+------------+-------+
|         80 |    56 |
+------------+-------+
|         70 |    58 |
+------------+-------+
|         60 |    60 |
+------------+-------+
|         50 |    62 |
+------------+-------+
|         40 |    64 |
+------------+-------+
|         30 |    66 |
+------------+-------+
|         20 |    69 |
+------------+-------+
|         10 |    73 |
+------------+-------+

  1. The generous people at Focused Objective have some more detailed examples and exercises available on their GitHub. ↩︎

  2. It’s a bit beyond the scope of this post, but there are methods you can use to determine the exact number of trials you need to run to achieve a particular confidence value in the outcome. ↩︎

  3. They probably mentally substituted the word ‘forecast’ for the word ‘commitment’ too. ↩︎