Professional Documents
Culture Documents
Case Study - SupChains - Forecasting
Case Study - SupChains - Forecasting
Executive Summary
In this two-part proof-of-concept, SupChains and DragonRitter (a forecasting-as-a-service
platform) helped a Pharma Distributor to reduce its inventory while optimizing its service
level. The POC showed a forecast error reduction of 25% and an expected inventory
excess reduction of 40% within six months while securing higher service levels (>95%
compared to the current 80%). In addition, the model provided the client with a list of dead
stock so they could take action immediately.
Based on these preliminary results (presented in November 2022), the client decided to use
our forecast engine as of December 2022 and plan its purchases based on the model as of
January 2023.
Business Situation
Our client is a pharma distributor active in Latin America with a dozen active warehouses and
around 10,000 unique products. In 2021, they reported a total revenue of 1B$, employing more
than a thousand employees.
As many supply chains worldwide, they suffer from both dead stocks and shortages: despite
stocking the equivalent of 45 days of sales, they only achieved a 60% fill rate in 2022 (up to
80% in the warehouse we used in the POC). They solicited the help of DragonRitter (a
forecasting-as-a-service platform) and SupChains to propose a joint end-to-end solution.
Our client’s objective is to increase their service levels while reducing their total inventory.
Forecast Challenges
To support its supply chain decisions, the distributor needs an accurate forecast per
warehouse, product, and week for the upcoming 26 weeks. This represents around 100,000
combinations. (The POC focused on a single warehouse.)
We have access to historical sales (per client channel) – which is the main requirement. But
the information fed into the model could be enriched by promotions, backorders, shortages,
nicolas.vandeput@supchains.com Confidential
Case Study: Using machine learning to forecast demand despite limited data
product information, prices, etc. Unfortunately, this project dataset presented two main
difficulties: erratic demand behavior and a lack of rich data. Let’s discuss these two in
detail.
Moreover, 99% of the sales observations are below 400 units, but the remaining 1% go up to
80,000 units sold in a single week.
Figure 2 99% of sales observations are below 400 units, but some products sold for up to 80,000 units in a single week.
To illustrate this erratic demand, you can see the anonymized demand pattern of one of their
best sellers in the figure below. This pattern is wildly irregular: the average order numbers
are four times higher than the median – and the demand coefficient of variation is around
150% (whereas, usually, top sellers enjoy more stable demand patterns).
nicolas.vandeput@supchains.com Confidential
Case Study: Using machine learning to forecast demand despite limited data
Figure 3 Even the top sellers display erratic demand patterns (anonymized data).
No Demand Drivers
Machine learning models are especially suited to gather insights from various demand
drivers (pricing, promotions, shortages) to generate accurate predictions.2
For this specific case, we could only gather historical prices.
1
We greatly encourage all supply chain practitioners to keep as much historical data as possible and
store it in a clean, consistent way. Applying data management best practices will show a return on
investment in forecasting accuracy, as shown in our previous case studies (here and here).
2
To see how our models deal with demand drivers such as promotions and pricing, see our previous
case studies here (manufacturer with promotions only) and here (retailer with promotions and
pricing).
nicolas.vandeput@supchains.com Confidential
Case Study: Using machine learning to forecast demand despite limited data
Figure 4 - Combination of the old and new versions of a similar product. The combined data will be used to forecast future
sales (data: chemical manufacturer).
SupChains Solution
Forecasting Metrics
The dataset being erratic, we have chosen a simple combination of Median Absolute Error
(MAE, or simply forecast error in the figures) and bias to assess the forecasting quality of our
nicolas.vandeput@supchains.com Confidential
Case Study: Using machine learning to forecast demand despite limited data
forecasts. This combination has the advantage of being simple to interpret while looking both
at accuracy and bias.3
Results
As illustrated hereunder, our machine learning models delivered more than 25% forecast
added value compared to a benchmark (a 12-weeks moving average).4
Figure 6 Our machine learning delivers an 27.9% FVA compared to the benchmark. The absolute error of the benchmark is
well above 100% due to the erratic demand patterns.
We could deliver this amazing 25% added value despite the lack of rich data. If the dataset
was enriched with demand drivers (marketing & promotions), life-cycle information
(predecessors and successors), and shortage data, we could achieve even better results.
3
More information about forecasting metrics here.
4
We compute the added value as the % reduction of the scoring metric (the combination of MAE and
Bias). More information about forecast value added here. For more info, see «How do we select our
benchmark ?» and «How do we test our models ?» at the end of the document.
5
This difference can be mechanically explained by the evolution of the bias: the absolute error is
usually reduced when the bias is lower. As the benchmark achieves a lower bias for long-term
predictions, it also benefits from a reduced absolute error.
nicolas.vandeput@supchains.com Confidential
Case Study: Using machine learning to forecast demand despite limited data
Project Timeline
As only little data cleaning was needed, only two weeks of work were needed to create the
model and test it. A final week was dedicated to analyzing the models’ performance and
creating the final report.
nicolas.vandeput@supchains.com Confidential