Forecasting of customer demands for production planning by local k-nearest neighbor models

https://doi.org/10.1016/j.ijpe.2020.107837Get rights and content

Highlights

  • Comprehensive empirical study utilizing a large dataset of industrial time series.

  • Low forecast errors and low computation times compared to established benchmarks.

  • High service levels and low safety stocks in an inventory simulation.

  • Best results among all applied methods regarding short time series.

Abstract

Demand forecasting is of major importance for manufacturing companies since it provides a basis for production planning. However, demand forecasting can be a difficult task because customer demands often fluctuate due to several influences. Methods of nonlinear dynamics have shown promising results in numerous applications but they have mostly been neglected in the context of demand forecasting. This paper evaluates the forecasting performance of local k-nearest neighbor models, which base on the theory of dynamical systems, in a comprehensive empirical study utilizing a large dataset of industrial time series of the M3-Competition. After a broad literature review, the theoretical background is described. Subsequently, different parameter configurations and model selection strategies are compared. A locally constant mean and a locally constant median are compared to locally linear regression models with four different regularization methods and different parameter configurations. In this comparison, the locally constant mean and the locally linear ridge regression with high regularization parameters provide the best trade-offs between forecast accuracy and computation times. Finally, these models achieve a high performance regarding low forecast errors, short computation times as well as high service levels in an inventory simulation compared to established benchmark methods. In particular, they obtain the best results among all applied methods regarding short time series. Moreover, they achieve the lowest errors considering the original accuracy criterion of the M3-Competition. Hence, local k-nearest neighbor models can be regarded as a valid alternative for demand forecasting in an industrial context, accomplishing high forecast accuracy with short computation times.

Introduction

Accurate demand forecasts are essential for all companies in a supply chain since several decisions have to be made before the actual values of important variables are known. In addition to forecasts of resource and energy demands (Hahn et al., 2009, Weinert et al., 2011, Hong and Fan, 2016), electricity prices (Aggarwal et al., 2009, Weron, 2014), failure prognosis for maintenance planning (Peng et al., 2010, Sikorska et al., 2011, Fritzsche et al., 2014, Gao et al., 2015) or yield and quality predictions (Lieber et al., 2013, Colledani et al., 2014), forecasts of future customer demands are necessary. In particular, accurate demand forecasts are crucial in build to stock productions, for instance often occurring in the fast moving consumer goods industry. In this context, raw material orders as well as production programs base strongly on demand forecasts and customer demands are directly satisfied from finished goods inventories. Normally, forecasts are calculated on a weekly or monthly basis for a horizon of approximately one year (Fleischmann et al., 2015) based on time series of past customer orders (Tempelmeier, 2008). Due to various influencing factors, demand evolutions are often highly volatile. For example, the demand of a customer depends on the customer’s remaining product inventory, the number of available substitute products on the market, the success of marketing campaigns, the current economic, political, social or ecological situation as well as basic agreements (Wiendahl et al., 2007, Porter, 2008). Therefore, forecasting future customer demands can be a difficult task.

A typical demand planning process in a company consists of six steps: (i) data acquisition and preparation, (ii) statistical forecasting, (iii) judgmental forecasting, (iv) consensus forecasting, (v) release of the forecast as well as (vi) evaluation and monitoring (Kilger and Wagner, 2015, Petropoulos et al., 2016). In this context, the step of statistical forecasting is of particular importance since it provides forecasts by automatic algorithms that are used as a basis for the following steps. Therefore, the paper at hand focuses on statistical forecasting and neglects the other steps of demand planning. It has to be noted that this paper considers the forecasting of regular customer demand characterized by time series with integer values greater than zero. For applications regarding intermittent demand, which is characterized by a sporadic demand evolution including periods without a demand, such as often observed for spare parts, see (Syntetos and Boylan, 2001, Wallström and Segerstedt, 2010, Kourentzes, 2013, Kourentzes, 2014, Van Wingerden et al., 2014).

Frequently, manufacturing companies have to forecast customer demands for thousands of products simultaneously. Hence, automatic forecasting methods have to be applied to calculate big amounts of forecasts. While simple forecasting methods, such as moving averages, simple exponential smoothing or linear regression methods are still often used in industrial practice (Küsters et al., 2006), more sophisticated methods already proved applicability in many practical applications as well as in forecasting competitions (Hyndman, 2020). Autoregressive integrated moving average (ARIMA) models (Hyndman and Khandakar, 2008, Babai et al., 2013, Box et al., 2015, Svetunkov and Boylan, 2019) and complex variants of exponential smoothing (Gardner, 2006, Hyndman et al., 2008, Ferbar Tratar et al., 2016, Kück et al., 2016b, Sbrana and Silvestrini, 2014, Sbrana and Silvestrini, 2019) are established methods that have shown potential in several studies, such as the M3-Competition (Makridakis and Hibon, 2000). In addition, machine learning methods, such as artificial neural networks (Crone et al., 2008, Crone and Kourentzes, 2010, Adeodato et al., 2011, Montavon et al., 2012, Kourentzes, 2013, Kourentzes et al., 2014) or support-vector machines (Cortes and Vapnik, 1995, Crone et al., 2008, Lu, 2014) have shown promising results in current studies, such as the NN3-Competition (Crone et al., 2011). However, these methods generally need long computation times and they carry the risk of overfitting when they are applied fully automatically. Moreover, some of these methods need long time series of past customer demands in order to fit their parameters in a suitable way. While demand forecasting can commonly be considered as a big data problem of calculating many forecasts for several products at regular intervals, it is nevertheless often the case that the individual demand histories only comprise a few years. Hence, computationally efficient methods that achieve a stable forecasting performance and do not need long demand histories to fit their parameters suitably would be beneficial.

Despite the variety of methods that have been applied to forecast industrial time series, local k-nearest neighbor models have been mostly neglected so far. Based on a modeling by nonlinear dynamical systems, these methods are able to reconstruct the dynamical properties of a demand evolution as well as influencing factors out of a scalar time series of past customer demands. These so-called methods of nonlinear dynamics have been successfully applied to conduct different tasks in production and logistics systems, such as modeling and control of production systems (Scholz-Reiter et al., 2002, Papakostas et al., 2009) or forecasting sporadic spare parts demand (Nikolopoulos et al., 2016). Furthermore, small-scale studies covering one-step ahead forecasts with small sets of exemplary customer demand time series have already shown promising results of local k-nearest neighbor forecasting models of nonlinear dynamics to forecast regular monthly customer demand (Mulhern and Caprara, 1994, Kück and Scholz-Reiter, 2013, Kück et al., 2014). However, the exemplary results of these studies can only be seen as an indication that these methods could perform well to forecast industrial time series and this assumption has to be proven in a comprehensive study.

The contributions of this paper are the following:

  • Proposition of local k-nearest neighbor models for regular monthly demand forecasting: The paper proposes local k-nearest neighbor models as a novel class of methods to forecast regular monthly customer demands of a manufacturing company.

  • Parameter comparison and model selection strategies: Two different versions of k-nearest neighbor models are compared, namely locally constant models and locally linear regression models. While standard parameters are used for delay coordinate embedding, different regularization methods with respective regularization parameters are compared regarding their suitability to forecast industrial time series. Moreover, different model selection strategies are proposed.

  • Evaluation regarding forecast accuracy: The forecasting performance of two pre-parameterized versions of the local k-nearest neighbor models as well as different individual model selection strategies are evaluated in a comprehensive empirical study utilizing 808 monthly industrial time series of the M3-Competition. As a first criterion of forecasting performance, the methods are evaluated in rolling-origin experiments with forecasts from multiple origins across short, medium and long horizons in comparison to several established benchmark methods. In addition to the evaluations in terms of mean forecast errors, Friedman and Nemenyi tests are conducted to assess significant differences between the performances of the local k-nearest neighbor models and the benchmark methods.

  • Evaluation regarding computation times: In contrast to the usual practice in other studies, this paper also evaluates the computation times of the forecasting methods. In this context, mean times per parameter optimization and per forecast are compared.

  • Evaluation regarding inventory performance: Another criterion that is often neglected in other studies is the inventory performance achieved based on calculated forecasts. This paper evaluates the local k-nearest neighbor models and the benchmark methods considering achieved service levels and needed safety stocks in an inventory simulation.

The remainder of this paper is structured as follows: Section 2 provides an overview of the existing literature regarding applications of local k-nearest neighbor forecasting models. Section 3 explains the theoretical background. After a parameter comparison and a comparison of different model selection strategies in Section 4, Section 5 comprises a comprehensive empirical study to assess the forecasting performance of the local k-nearest neighbor models in terms of forecast errors and computation times. Section 6 applies the forecasting methods in an inventory simulation and evaluates their performance regarding inventory metrics. Section 7 concludes the paper.

Section snippets

Literature overview

Local k-nearest neighbor models belong to the methods of nonlinear dynamics, which originate from the theory of dynamical systems and which base on the paradigm of deterministic chaos. For general overviews regarding methods of nonlinear dynamics, see Kantz and Schreiber, 2004, Aguirre and Letellier, 2009 and Broer and Takens (2011). Apart from successful applications in modeling, analysis and control of manufacturing systems (Ramadge, 1993, Prabhu and Duffie, 1995, Wiendahl and Scheffczyk, 1999

Theoretical background

This section describes the necessary steps to build a local k-nearest neighbor (knn) forecasting model. After an introductory overview, the steps of delay coordinate embedding, knn searching and building a local knn forecasting model are detailed.

Parameter comparison and model selection strategies

This section determines valuable parameter configurations and model selection strategies for the local knn models to forecast monthly customer demands. The section starts with an experiment description, considering the compared parameter configurations, the used dataset and the performance criteria, which are mean forecast accuracies and computation times. Afterwards, the results of the parameter comparison are shown and suitable model selection strategies are determined.

Empirical study 1: Forecast errors and computation times

This section evaluates the performance of the four model selection strategies for local knn forecasting models determined in the last section. After an experimental description, the section comprises evaluations regarding forecast errors as well as times per forecast and per optimization.

Empirical study 2: Inventory simulation

After evaluating the performance of the forecasting methods in terms of forecast accuracy and computation times, this section evaluates their performance in inventory simulations. For this purpose, a discrete-event simulation model has been implemented. For a given time series of customer demands, each of the forecasting methods is applied to forecast the demands and a periodic order-up-to policy is used to calculate product replenishment orders. Section 6.1 describes the simulation model and

Conclusion

This paper has assessed the suitability of local k-nearest neighbor models to forecast industrial time series data. After a thorough description of the theoretical background, an extensive parameter comparison has been conducted to find reasonable parameters and model selection strategies that lead to stable forecasting performance. In this context, the locally constant knn model was tested with a local averaging according to an arithmetic mean and a median. Moreover, different regularization

CRediT authorship contribution statement

Mirko Kück: Conceptualization, Methodology, Software, Validation, Formal analysis, Investigation, Writing - original draft, Writing - review & editing, Visualization. Michael Freitag: Conceptualization, Writing - review & editing, Supervision, Project administration.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References (140)

  • CroneS.F. et al.

    Feature selection for time series prediction - A combined filter and wrapper approach for neural networks

    Neurocomputing

    (2010)
  • DonnerR. et al.

    Nonlinear characterization of the performance of production and logistics networks

    J. Manuf. Syst.

    (2008)
  • Ferbar TratarL. et al.

    Demand forecasting with four-parameter exponential smoothing

    Int. J. Prod. Econ.

    (2016)
  • Fernández-RodríguezF. et al.

    Exchange-rate forecasts with simultaneous nearest-neighbour methods: evidence from the EMS

    Int. J. Forecast.

    (1999)
  • FreitagM. et al.

    Dynamics of resource sharing in production networks

    CIRP Ann. Manuf. Technol.

    (2015)
  • FritzscheR. et al.

    Optimal prognostic distance to minimize total maintenance cost: The case of the airline industry

    Int. J. Prod. Econ.

    (2014)
  • GaoR. et al.

    Cloud-enabled prognosis for manufacturing

    CIRP Ann. Manuf. Technol.

    (2015)
  • GardnerE.S.

    Exponential smoothing: The state of the art-Part II

    Int. J. Forecast.

    (2006)
  • HahnH. et al.

    Electric load forecasting methods: Tools for decision making

    European J. Oper. Res.

    (2009)
  • HongT. et al.

    Probabilistic electric load forecasting: A tutorial review

    Int. J. Forecast.

    (2016)
  • HwarngH.B. et al.

    Understanding supply chain dynamics: A chaos perspective

    European J. Oper. Res.

    (2008)
  • HyndmanR.J.

    A brief history of forecasting competitions

    Int. J. Forecast.

    (2020)
  • JayawardenaA.W. et al.

    Neighbourhood selection for local modelling and prediction of hydrological time series

    J. Hydrol.

    (2002)
  • JursaR. et al.

    Short-term wind power forecasting using evolutionary algorithms for the automated specification of artificial intelligence models

    Int. J. Forecast.

    (2008)
  • KantzH. et al.

    Markov Chain model for turbulent wind speed data

    Physica A

    (2004)
  • KourentzesN.

    Intermittent demand forecasts with neural networks

    Int. J. Prod. Econ.

    (2013)
  • KourentzesN.

    On intermittent demand model optimisation and selection

    Int. J. Prod. Econ.

    (2014)
  • KourentzesN. et al.

    Neural network ensemble operators for time series forecasting

    Expert Syst. Appl.

    (2014)
  • KückM. et al.

    Emergence of non-predictable dynamics caused by shared resources in production networks

    Procedia CIRP

    (2016)
  • KückM. et al.

    Robust methods for the prediction of customer demands based on nonlinear dynamical systems

    Procedia CIRP

    (2014)
  • KugiumtzisD. et al.

    Regularized local linear prediction of chaotic time series

    Physica D

    (1998)
  • KüstersU. et al.

    Forecasting software: Past, present and future

    Int. J. Forecast.

    (2006)
  • LeeY.S.

    A semi-parametric approach for estimating critical fractiles under autocorrelated demand

    European J. Oper. Res.

    (2014)
  • LieberD. et al.

    Quality prediction in interlinked manufacturing processes based on supervised & unsupervised machine learning

    Procedia CIRP

    (2013)
  • LisiF. et al.

    Is a random walk the best exchange rate predictor?

    Int. J. Forecast.

    (1997)
  • LisiF. et al.

    A comparison between neural networks and chaotic models for exchange rate prediction

    Comput. Statist. Data Anal.

    (1999)
  • LuC.-J.

    Sales forecasting of computer products based on variable selection scheme and support vector regression

    Neurocomputing

    (2014)
  • MakridakisS. et al.

    The M3-competition: Results, conclusions and implications

    Int. J. Forecast.

    (2000)
  • McNamesJ.

    Local averaging optimization for chaotic time series prediction

    Neurocomputing

    (2002)
  • MeadeN.

    A comparison of the accuracy of short term foreign exchange forecasting methods

    Int. J. Forecast.

    (2002)
  • MinnerS.

    Multiple-supplier inventory models in supply chain management: A review

    Int. J. Prod. Econ.

    (2003)
  • MulhernF.J. et al.

    A nearest neighbor model for forecasting market response

    Int. J. Forecast.

    (1994)
  • NikolopoulosK.I. et al.

    Forecasting supply chain sporadic demand with nearest neighbor approaches

    Int. J. Prod. Econ.

    (2016)
  • NikolopoulosK. et al.

    Forecasting with cue information: A comparison of multiple regression with alternative forecasting approaches

    European J. Oper. Res.

    (2007)
  • PapakostasN. et al.

    Modelling the complexity of manufacturing systems using nonlinear dynamics approaches

    CIRP Ann. Manuf. Technol.

    (2009)
  • PetropoulosF. et al.

    Do big losses in judgmental adjustments to statistical forecasts affect experts behaviour?

    European J. Oper. Res.

    (2016)
  • PetropoulosF. et al.

    The inventory performance of forecasting methods: Evidence from the M3 competition data

    Int. J. Forecast.

    (2019)
  • PrabhuV.V. et al.

    Modelling and analysis of nonlinear dynamics in autonomous heterarchical manufacturing systems control

    CIRP Ann. Manuf. Technol.

    (1995)
  • AbarbanelH.D.I. et al.

    Prediction in chaotic nonlinear systems: Methods for time series with broadband Fourier spectra

    Phys. Rev. A

    (1990)
  • AguirreL.A. et al.

    Modeling nonlinear dynamics and chaos: A review

    Math. Probl. Eng.

    (2009)
  • Cited by (56)

    • Integrated lot-sizing and scheduling: Mitigation of uncertainty in demand and processing time by machine learning

      2023, Engineering Applications of Artificial Intelligence
      Citation Excerpt :

      We consider three factors influencing the demand as follows: Generally, time-varying features are widely used as well-known inputs in the prediction of the demand Kück and Freitag (2021). Therefore, we consider the dependency between the amount of demand and the period’s number (e.g., seasonal changes) as the first factor.

    View all citing articles on Scopus
    View full text