Applying ML... Accepted Version
Applying ML... Accepted Version
Applying ML... Accepted Version
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
conditions evolve over time. Many factors intervene simultaneously and their hard-to-
interpret interactions throughout the supply chain greatly complicate decision making.
The complexity clearly manifests itself in the field of inventory management, in which
This paper applies machine learning to help managers understand these complex
scenarios and better manage the inventory flow. Building on a dynamic framework,
defined by seven variables (cost structure, demand variability, three lead times, and
determines the best replenishment rule around 88% of the time. This leads to a
observe that the nodes are much more sensitive to inventory decisions in the lower
1. Introduction
Globalization has utterly changed the business landscape, where competition has
not only increased substantially but also become more complex and dynamic (Puche et al.
2016). This competition has indeed moved from the firm level to the network level, placing
1
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
(Melnyk et al. 2009). However, these advantages are difficult to capture. Managers must
convoluted supply networks with long and variable lead times, and be able to agilely react
to the frequent changes in the environment (Mentzer et al. 2001). Comprehending the
supply chain interdependencies between processes, decisions, and structures is far from
The complexity becomes evident in the field of inventory management, one of the
cornerstones of the supply chain discipline. APICS (2011, 48) defines inventory as “an
expensive asset” that “needs to be carefully managed”, whose primary purpose is “to meet
evaluate two primary aspects when making replenishment decisions to control the
inventory flow (Disney and Lambrecht 2008). First, they must consider a key trade-off
between inventory investment and service level, with the aim of satisfying consumer
demand in a cost-effective manner (Steinker, Pesch, and Hoberg 2016). Second, they need
variability of production schedules and hence may trigger different sources of costs, e.g.
extra capacity, overtime, and idle time (Disney et al. 2006). Overall, Lancioni (2000)
claimed that inventory-related costs cover nearly 50% of the supply chain costs.
the performance of supply chains. To this end, managers need to consider the impact of the
complex interactions between a wide range of variables, which may result in an intractable
problem (Bischak et al. 2014). This task becomes even more difficult in what we label as
environment (e.g. consumer demand, raw materials cost, or stakeholders’ decisions) suffer
2
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
from frequent changes over time (Chopra and Sodhi 2004). In these cases, it may be
From this perspective, this work develops a dynamic framework for managing
inventories in the supply chain. The framework employs machine learning, specifically
inductive learning, for understanding the complex relationships between the controllable
and uncontrollable factors that impact on business performance. It has been designed to
periodically select the best inventory policy, among a set of baseline rules, according to the
show that machine learning can help managers make decisions that are hard to deal with
from other approaches, which eventually would result in an increased performance. In this
sense, machine learning techniques may be interpreted as a promising next step in the field
of inventory management.
problem —we focus on the measurement of performance and present some established
replenishment policies. Section 3 introduces the inductive algorithm that we use and delves
describes the dynamic framework we propose for managing inventories. Section 5 presents
the case study where we test our proposal, and details the generation of examples for the
learning algorithm. Section 6 shows the numerical results and evaluate them against the
static alternative. Finally, Section 7 concludes and reflects on the implications of this
research.
3
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
powerful enemy: the Bullwhip Effect (Lee, Padmanabhan, and Whang 1997). This
phenomenon is common in all kinds of industries (see e.g. Isaksson and Seifert 2006) and
may reduce the profitability of firms significantly (Metters 1997). It refers to the tendency
of the variability of the signals, mainly orders and consequently inventories, to increase as
they pass through the various nodes of the supply chain; see the recent review by Wang
and Disney (2016) for further detail. From the previous definition, two ratios are
commonly used to quantify the Bullwhip Effect: the order variance ratio (OVR) and the
inventory variance ratio (IVR). The former compares the variance of the orders issued (𝜎𝑂2 )
and received by the node, i.e. its demand (𝜎𝐷2 ), by eq. (1); while the latter quantifies the
2
variance of the net stock1 (𝜎𝑁𝑆 ) against the demand variability, by eq. (2).
𝜎𝑂2
𝑂𝑉𝑅 = (1)
𝜎𝐷2
2
𝜎𝑁𝑆
𝐼𝑉𝑅 = 2 (2)
𝜎𝐷
As previously discussed, decision makers need to consider both the production and
metrics cover both aspects; thus defining a powerful framework for evaluating the
operational performance of supply chains (Cannella et al. 2013). First, OVR measures
schedules that significantly decrease supply chain efficiency. Indeed, Disney, Gaalman,
and Hosoda (2012) showed that the minimum production cost is proportional to the square
4
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
root of OVR in linear guaranteed-capacity models2. Second, IVR considers net stock
variability, which determines the firm’s ability to meet effectively a predetermined service
level. Hence, reducing IVR is essential to appropriately balance the risk of breaking stock
and the charge of holding too much stock. In this sense, Kahn (1987) showed that the
minimum inventory cost is linearly related to the square root of the net stock (and thus
IVR) when holding and backlog costs are proportional to the volume.
The function 𝐽 fuses both indicators into one metric, through a weighted sum of
their square roots; see eq. (3). Here 𝑤𝑜 and 𝑤𝑖 (𝑤𝑜 , 𝑤𝑖 ≥ 0, 𝑤𝑜 + 𝑤𝑖 = 1) depend on the
cost associated to each source of variability and express the relative importance of each
indicator. For example, 𝑤𝑜 = 0.8 (𝑤𝑖 = 0.2) would reveal that order variability is more
damaging; while 𝑤𝑜 = 0.2 (𝑤𝑖 = 0.8) would illustrate the opposite scenario. Following
from the previous discussion, it can be assumed that 𝐽 provides a fair understanding of the
cost performance of a determined inventory police. For this reason, we employ this metric
in this work. For further details on 𝐽, please refer to Ponte, Wang et al. (2017).
To control the inventory, there are several types of replenishment strategies (see
e.g. Zipkin 2000). This paper is concerned with the order-up-to (OUT) family, which
review inventories and place orders at fixed intervals. These periodic-review systems are
systems (Axsäter 2003). They also produce benefits from other perspectives; for instance,
they enable combined orders to save transportation costs (APICS 2011). Hence, it is a
5
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
(Sillanpää and Liesiö 2018) and OUT policies are widely used in real supply chains
OUT policies place orders periodically, e.g. at the end of each period t, to bring the
inventory up to a determined level. The traditional OUT model (e.g. Disney and Lambrecht
gaps, by eq. (4). First, between the safety stock (𝑆𝑆𝑡 ) and the actual net stock (𝑁𝑆𝑡 ); and
second, between the desired and the actual work-in-progress (𝐷𝑊𝑡 , 𝐴𝑊𝑡 ). Note that work-
in-progress covers the product that has been ordered but not yet received.
When the safety stock is appropriately adjusted, the OUT model finds the optimal
balance between holding and backlog costs (Karlin 1960). In this sense, this policy is able
to minimize the IVR metric. Nevertheless, it generally offers poor performance from the
that the OVR generated by this policy is always greater than 1 for three common
forecasting methods. To sum up, Gaalman (2006, 1284) states that the OUT policy “will
mainly minimize inventory costs or equivalently inventory variance”, but “the control of
(0 ≤ 𝛽 ≤ 1) into the ordering rule to regulate the amount of gaps to be recovered; e.g. Lin
et al. (2017) reviews several applications of inventory controllers over the last decades.
This results in the so-called proportional order-up-to (POUT) policy, see eq. (5).
Depending on the value of the controller, this policy allows modeling a wide range of real-
world replenishment strategies (Li and Disney 2017). The smaller 𝛽, the less sensitive the
order to the inventory gaps. This simple mechanism allows to directly control, and reduce,
6
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
order variability (Disney and Lambrecht 2008). Gaalman (2006) concluded that the POUT
reduced, OVR tends to decrease at the expense of an increase in IVR; e.g. see Figure 1 in
Ponte, Sierra et al. (2017). Therefore, reducing 𝛽 allows managers to decrease ordering-
related costs, generally at the expense of increasing inventory-related costs. In light of this,
the tuning of the controller has become a fruitful area of study with the aim of finding the
right balance between both metrics; see e.g. Cannella and Ciancimino (2010). However,
very high, generally being an intractable problem through analytical techniques (see e.g.
Disney et al. 2006). In this paper, we consider the impact of a wide range of uncontrollable
and controllable factors, and their interplays, on determining a suitable value for the
controller.
development of algorithms capable of learning from data. These techniques can be applied
to solve different kinds of problems using knowledge obtained from similar past problems
(Michalski, Carbonell, and Mitchell 1983). According to the review by Priore et al. (2014),
the main machine learning techniques are: (1) inductive learning; (2) artificial neural
networks; (3) case-based reasoning; (4) support vector machines; and (5) reinforcement
7
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
results in a set of decision rules that build a decision tree. Thence, this conceptual approach
allows users to easily understand the decision-making process (Filipic and Junkar 2000).
The learning algorithm obtains the knowledge by examining a training dataset. This
includes the past problems and their solutions (examples) and can be represented as an
attribute-value table. The input attributes refer to the features of the problem, while a
special attribute named “class” includes the optimal solution. Inductive learning techniques
recursively split this initial dataset into subsets depending on the value of one attribute.
This results in the generation of the decision tree, which is employed to solve new
problems by assigning a class to the set of values of the attributes defining them. Note that
information about the solved problems may thus be used to analyze future problems. In
this sense, this approach incorporates principles of information updating, which is gaining
interest as an important process for supply chain learning (Shen, Choi, and Minner 2018).
From the pioneering works by Hoveland and Hunt in the 1950s, there is a wide
range of inductive learning algorithms. The CART (Friedman 1977), ID3 (Quinlan 1979),
PLS (Rendell 1983), ASSISTANT 86 (Cestnik, Kononenko, and Bratko 1987), and C4.5
(Quinlan 1993) deserve to be mentioned here. The last one is generally considered the
most popular inductive learning algorithm (Wu et al. 2008; Witten et al. 2016), as it can
achieve a very good trade-off between error rate and speed of learning (Lim, Loh, and Shih
2000). For this reason, we employ this algorithm in this research work.
The C4.5 algorithm uses the concept of information entropy to sequentially select
the nodes of the tree. This refers to the amount of information produced by a source of data
8
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
and can be formally expressed by eq. (6) for a set 𝐷 of cases, where 𝐶 denotes the number
of classes. Note 𝑝(𝐷, 𝑗) refers to the proportion of cases in 𝐷 that belong to the j-th class,
2008). First, it checks if either all the cases in the dataset 𝑆 belong to the same class or 𝑆 is
small. If so, it simply creates a leaf node for the tree with the most frequent class.
Otherwise, it calculates the information gain (the change in information entropy compared
to the previous state) from splitting on each attribute 𝐴𝑋 and creates a node based on the
attribute that maximizes the information gain. This can be maximized in absolute terms
(𝑔1 ) or in relative terms to the information provided by the test sources (𝑔2 , which corrects
the gain by considering information about the class)3. Then, it recurs on the obtained
subsets through the same procedure. Last, the tree is pruned from the leaves to the root to
avoid overfitting. We refer the interested reader to Wu et al. (2008) for more details on the
pruning algorithm.
the complexity and dynamism of the current business scene. Accordingly, practitioners and
academics have explored ways to better manage the information and leverage this to make
more robust decisions; e.g. see the review by Ko, Tiwari, and Mehnen (2010). In line with
the previous discussion, machine learning can be of special interest in this regard. Next, we
review the relevant literature that applies these techniques to the control of inventories in
the supply chain. These studies represent the background of our research work.
9
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
inventory at all nodes of the supply chain in a coordinated manner, such as Giannoccaro
and Pontrandolfo (2002), Chaharsooghi, Heydari, and Zegordi (2008), and Mortazavi,
Khamseh, and Azimi (2015). Their solutions employ different algorithms for
reinforcement learning, e.g. Q-learning (Watkins and Dayan 1992), to determine near-
optimal ordering policies. To this end, they use simulation techniques to explore the
behavior of the supply chain in a wide range of scenarios. The proposed solution takes
decisions according to the system state vector, which is generally defined as formed by the
inventory position of the various supply chain nodes. In these works, the learning-based
A slightly different approach is that by Sui, Gosavi, and Lin (2010) and Akhbari et
al. (2014), both focusing on vendor-managed inventory systems. The former employ
reinforcement learning for determining the optimal retailer’s replenishment policy. Their
solution, considering two products, also calculates the number of trucks dispatched by a
distribution center to a set of retailers. The latter concentrate on determining the optimal
production policy for the manufacturer. They use case-based reasoning by means of the
continuous K-nearest neighbor algorithm. Both articles show that the learning-based
approach effectively increases the profit of the supply chain over traditional methods.
The usefulness of machine learning for managing the inventory flow through an
automatic configuration of the supply chain has also been investigated. Piramuthu (2005a)
develops an inductive learning-based tool that determines dynamically the optimal supplier
for the different nodes depending on the lead times and the order quantity. Piramuthu
(2005b) extends this framework to a multi-product context. In both cases, this dynamic
10
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Last, several authors explore the effectiveness of these techniques for demand
Carbonneau, Laframboise, and Vahidov (2008) show that recurrent neural networks and
support vector machines are able to provide very accurate forecasts for real-world datasets,
resulting in an improved inventory control. Several recent works follow this research line,
see e.g. the reviews by Bajari et al. (2015) and Syam and Sharma (2018).
In line with previous works (e.g. Min 2010, Kuo and Kusiak 2018), we conclude
that despite its widespread acceptance as a tool for improving decision-making processes,
the applications of machine learning are yet emerging in the supply chain field. There is a
wide range of processes that may strongly benefit from the use of these techniques, which
would result in strong competitive advantages for firms. It should be highlighted that one
of the main advantages of these artificial intelligence techniques is their dynamic nature
(Syam and Sharma 2018). This makes them especially suitable for a business scene like the
Our work combines ideas from the above avenues of research but follows a
framework for setting the most appropriate replenishment policy over time in dynamic
a wide range of both internal and external factors, as opposed to previous works in this
field. Despite the existence of more advanced algorithms, we use inductive learning as it
factors; unlike most machine learning techniques, which are generally considered “black-
11
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
inventory implications in different settings; e.g. see Khouja (1999) for a review in the
implications of replenishment policies are also considered (Disney et al. 2006); therefore,
Several methodologies, such as control theory (e.g. Lin, Spiegler, and Naim 2018) and
simulation (e.g. Cannella and Ciancimino 2010), have successfully helped to understand
the behavior of different policies; however, the question of optimality have been barely
previously discussed, they can enable managers to interpret complex interdependences and
provide near-optimal solutions to this problem; thus suggesting an interesting avenue for
In light of this, our approach is built on the dynamic framework for automated
inventory model for a node of the supply chain not only according to its state, but also
considering the state of its environment. In this sense, this control system is designed to
understand the multiple variables, both internal and external, impacting on the node’s
performance and construct a decision tree that governs the inventory flow. By altering the
ordering policy depending on the context where the node operates, we expect to improve
12
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
attributes) together with the best inventory policy (class) in this scenario. In the continuous
operation of the system, examples for the training set may be obtained from refining the
accumulated feedback on its state and performance. However, creating a large mass
examples this way may be a very long process. This emphasizes the usefulness of a
simulation model that replicates the known environment for populating the example
dataset. Through this dataset, the inductive learning algorithm can be capable of acquiring
The decision tree acts as the regulator of the inventory management system,
establishing the replenishment model according to the firm’s and the supply chain state
over time. Dashed lines in Figure 2 underscore the key role of the supply chain
environment in this process, which interacts with the firm in a double way. On the one
hand, the supply chain greatly affects the firm’s performance —thereby, these factors must
also be considered by the control system. On the other hand, the node’s decisions impact
on its supply chain partners, which creates a hard-to-interpret loop. Considering this
This generic framework can be applied to any kind of supply chain from a single-
echelon perspective. No assumptions have been taken on the nature of the supply chain.
make a difference in the previously defined fast-changing environments, where the values
of the relevant variables rapidly evolve over time. In highly static environments, it may not
Finally, we would like to note that three aspects must be taken into serious
13
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
depends on the attributes; therefore, the key factors must be carefully selected and
appropriately measured. Second, achieving a large enough example set is essential to avoid
inadequate generalizations that reduce the efficiency of the system. Last, modifying the
inventory policy may generate an instable transitory (i.e. changing the policy too
frequently may result in poor system performance); therefore, the review period of the
supply chain that plays a key role in the distribution of a specific product. This node,
labelled as the wholesaler, purchases said product from a factory, which manufactures the
product, and later sells it to a retailer, which is the one directly dealing with the consumer.
We thus study a single-product serial supply chain composed of three nodes, see Figure 3.
The downstream material flow —from the factory to the consumer— comprises
three fixed lead times: one production lead time, associated to the manufacturing (𝑇𝑓 ), and
two shipping lead times, covering the transportation between nodes (𝑇𝑤 , 𝑇𝑟 ). The upstream
information flow —orders travel in the opposite direction— is triggered by the consumer
demand. This is considered to follow a normal distribution 𝑁(𝜇, 𝜎 2 ), where the coefficient
14
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
An important assumption behind our supply chain model is that the three nodes
following four-step sequence of events (per period, which we understand as a day) for the
discrete operation of these nodes, which is common in this kind of studies (e.g. Disney et
al. 2016). We do not include the mathematical formulation of the model in full detail due
to length restrictions and given that these difference equations are well known in the
problem-specific literature.
1. Reception state. The product is received (corresponding to the order placed before
the relevant lead time) and added to the net stock, and the order is observed. We
2. Serving state. The order received and past backorders (if they exist) are met from
net stock. Then, the product is sent downstream. We do not consider defective
3. Updating state. The inventory positions (both net stock and work-in-progress) are
updated and, if necessary, a backorder is generated. Note that these are allowed,
and the product will be delivered as soon as net stock becomes available.
4. Sourcing state. The order is issued according to a POUT policy. We assume the
quantity cannot be negative, i.e. excess products cannot be returned to the supplier.
POUT models, as per the previous description (in Section 2.2), incorporate four
decision points: controller setting, safety stock, forecast, and work-in-progress policy. We
distributed demands represent minimum mean square error (MMSE) forecasts (Disney et
al. 2016). Regarding the work-in-progress policy, we use the common solution 𝐷𝑊𝑡 = 𝑇𝑥 𝜇
long-term drift in the inventory position (Disney and Towill 2005). Besides, we consider
15
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
that the safety stock factor is 3, i.e. 𝑆𝑆𝑡 = 3𝜇, in line with prior works in the literature (e.g.
Ciancimino et al. 2012). Thus, we focus on the proportional controllers as the main
Finally, we would like to note that this supply chain model has several sources of
complexity, e.g. multi-echelon (Ciancimino et al. 2012) and nonlinear effects (Ponte,
Wang et al. 2017), which bring it closer to real-world environments but make that
Besides, we would like to underline that we use a generic, instead of specific, supply chain
conclusions.
The example generator is aimed at providing the machine learning algorithm with
the necessary information so that it is able to determine the best inventory policy for the
wholesaler in each possible scenario. Thus, the class of the examples refer to this optimal
policy. In this regard, we model four different policies: (1) OUT represents the classic
OUT model (i.e. βw=1); (2) POUT_H refers to a POUT model whose controller is
regulated at a high level (we select βw=0.7); (3) POUT_M represents a POUT model whose
controller is set at a moderate level (we select βw=0.4); and (4) POUT_L refers to a POUT
attributes to be representative of the node’s state and its environment: (1) the coefficient of
variation of the demand (CV), which ranges between 10% and 50%; (2) the three lead
times (𝑇𝑟 , 𝑇𝑤 , 𝑇𝑓 ), which vary between 1 and 4 days, (3) the setting of the retailer’s and
factory’s controller (𝛽𝑟 , 𝛽𝑓 ), which are randomly generated in the interval [0,1]; and (4) the
16
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
operation of the example generator are described in Figure 4. After randomly creating the
values of the seven attributes, the same scenario is run for the four policies in the
wholesaler, which requires previously initializing the system. Each run consists of 20,000
days —a large enough interval to ensure the stability of the response. After the four runs,
the class is selected as the policy that obtains the lowest value of the metric 𝐽. This
generates one example, and the process is repeated until obtaining 2,000 examples. To
To obtain the inventory management knowledge from the training dataset and
structure it through a decision tree, we employ the C4.5 algorithm in the data science
software RapidMiner. We use the cross-validation method to validate the results. This
randomly divides the example set into ten different blocks, nine of which are employed to
obtain the knowledge. The remaining one is used to test the decision tree by calculating the
number of examples appropriately classified. We repeat this process ten times and we
average the results, which defines the so-called hit ratio. This metric reports on the
17
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
accuracy of the inductive learning algorithm. Figure 5 displays the hit ratio for different
As expected, the hit ratio increases as the number of examples grows. Nonetheless,
this indicator stabilizes in a narrow range, approx. 87%-89%, over 600 examples. The
slight variability would then be mainly explained by the randomness of the examples
chosen to validate the algorithm. Overall, we observe that the proposed knowledge-based
system is capable of capturing the complex relationships between the different internal and
external factors that impact on supply chain performance, determining in approx. 8 out of
each 9 scenarios the best replenishment policy for the considered node.
In this and the next subsection, we consider the knowledge-based control system
obtained for 2,000 examples. This contains the most information on the attributes, with the
knowledge being structured around 88 decision rules including the seven attributes. By
way of illustration, Table 2 reports some of these rules. After each rule, we show the
number of examples of the dataset that are properly classified over the total number of
These 88 rules shape a complex decision tree. For the sake of clarity, we only
represent a simplified version of the tree in Figure 6. This shows the branches generated
from the two upper variables, respectively, the cost structure of the node represented by 𝑤𝑜
and the retailer’s inventory controller 𝛽𝑟 . At the bottom of this graph, we include the
18
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
replenishment policies in which each branch ends. Selecting among the different policies
A major notion derived from the decision tree is the order of relevance of the
factors. The tree underscores the weight 𝑤𝑜 as the most relevant one. This is interesting but
not surprising. It is well known that the optimal value of the inventory controller greatly
depends on the cost structure of the node. More unexpected is the finding that the
replenishment policy of the retailer (through its controller) is the second factor in terms of
importance. This reveals that the ordering policy of the lower echelon of the supply chain
greatly impacts on the optimal policy of the wholesaler. Given that the factory’s inventory
controller is placed much lower in the tree, we interestingly observe that the optimal
ordering rule of the wholesaler is more sensitive to the inventory decisions in the lower
nodes of the supply chain than to those in the upper nodes. The effect of the different lead
times and the demand variability is also less significant than that of the previous attributes.
Moreover, the decision tree allows decision makers to understand the cause-effect
relationships between the value of the attributes and their optimal policies. In this regard,
Figure 6 shows that when 𝑤𝑜 ≤ 0.748, the inventory controller should never be regulated
at low level; while when 𝑤𝑜 > 0.748, the controller should only be regulated at low or
medium level (unless 𝛽𝑟 is extremely low). Thus, the more relevant the production costs
compared to the inventory costs (i.e. the higher 𝑤𝑜 ), the stronger the node’s motivation to
regulate the inventory controller at low levels. Similarly, when 𝛽𝑟 is low —and hence the
orders issued by the retailer are relatively stable, thus mitigating the Bullwhip Effect—, the
wholesaler should opt for high values of the controller. However, when 𝛽𝑟 is high —the
retailer contributes to amplifying order variability—, the wholesaler should select low
19
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
values of the controller —which mitigates Bullwhip. For example, if 𝑤𝑜 = 0.8, the
wholesaler should employ an OUT policy or a POUT policy regulated at high level
(depending on the other attributes) when 𝛽𝑟 = 0.2, but this node should use a POUT policy
We now compare the performance of the supply chain operating with the dynamic
framework we propose with the static alternative. To this end, we run several simulation
runs of 500 months of 30 days. In the static case, the same inventory policy is always
employed over time. Meanwhile, in the dynamic framework, we consider that the
wholesaler evaluates its internal and external conditions at the beginning of each month, it
selects the optimal replenishment policy, and it operates with this policy until next month.
That is, the review period of the dynamic framework is set as 30 days.
considering the dynamism of the current business scene. From this perspective, we
evaluate its performance in two different scenarios. In the first one, labelled as fast-
changing scenario, the system randomly creates an initial combination of attributes at the
moderately modifying the previous values: within the interval ±1 for the (three) discrete
lead times and ±10% for the (four) continuous attributes. In the second one, labelled as
chaotic scenario, the values of the attributes are randomly generated each month; hence,
the attributes may dramatically change from one month to the next one.
Table 3 displays the results of three simulation runs in the fast-changing scenario.
In line with previous discussions, we measure operational performance through the average
value of the metric 𝐽, which is a proxy indicator of the sum of the inventory and production
20
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
costs incurred by the node. The first four rows show the results of the four policies if they
were used statically throughout the whole simulation horizon. The sixth row shows the
solution provided by our dynamic approach. For the sake of readability, the values in the
table are relative to the lowest possible 𝐽 (fifth row). This value (1.000), representing the
target for each simulation run, would be obtained if the inductive learning algorithm was
always capable of selecting the best policy (i.e. hit ratio = 100%).
Table 3 provides evidence of how the static approach generates a wide range of
average 𝐽 between 18.9% (run 2) and 21.9% (run 3) higher than the target (1.000). These
results reveal that the one-shot configuration may be inappropriate in scenarios which
undergo significant changes over time. At the same time, Table 3 illustrates that the
increase between 4.5% (run 1) and 6.7% (run 2) in 𝐽, thus dramatically outperforming the
use of the best policy from a static perspective. In light of this, the dynamic adjustment of
the inventory policy in response to the changes in the environmental conditions can
Table 4 presents the results for the chaotic scenario. In this case, the difference
between the best static policy and the dynamic solution grows. While the avoidable costs
generated by the former increase (the lowest 𝐽 in the static approach is now around 25%
higher than the optimal), those generated by knowledge-based framework are similar as
before (the increase in 𝐽 is slightly above 5%). Note that in this scenario the results vary
less between the three runs than before. Similarly, the best static policy here is the
POUT_H in the three runs, while in the previous case it was different in each run. This
21
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
occurs because the results of the fast-changing scenario are much more sensitive to the
randomly generated starting point (environmental conditions in each month depend on the
framework outperforms the best static decision through ANOVA techniques. We have
tested the significance of the difference between the means of both alternatives, and we
have obtained a p-value much lower than 5%. Thus, we reject the null hypothesis (equality
All in all, our results show how real-world businesses may suffer from their
inventory strategies becoming obsolete due to the evolving nature of the current business
scene. That is, a specific replenishment rule may work well at a certain point in time (i.e. in
demand uncertainty increases, or if retailers change their inventory policies). From this
perspective, we have observed the operational benefits derived from adapting the
relationships between the environmental factors and the optimal policy may become an
inextricable problem. In this regard, we demonstrate that the use of machine learning
22
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
such as those derived from stockouts, holding too much inventory, and unstable production
in this rapidly changing business scenario, one-shot approaches may not be enough, and
companies may benefit from rethinking the suitability of their inventory policy over time.
The present study approaches this problem by proposing a dynamic framework for
periodically determining the best replenishment rule for a specific supply chain node. This
has been designed to consider both internal and external factors, which constitutes a
relevant difference from prior works. Artificial intelligence methods are the backbone of
this framework. They can help decision makers to elucidate such a complex problem,
The first step for practitioners wishing to implement this dynamic approach would
simulation model. This process includes capturing the key variables that impact on
operational performance. The model would allow one to explore a wide range of scenarios
and investigate the suitability of each inventory policy in them. This information can then
be translated into knowledge by a machine learning algorithm, which could establish a set
of decision rules for the control of the real-world system over time; thus, equipping firms
algorithm has proven to successfully deal with the convoluted nature of a seven-variable
inventory management problem, selecting (among four alternatives) the best inventory
policy for a wholesaler with an average accuracy of 88%. This results in a significant
reduction of the operating costs in comparison with the best static alternative. The
23
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
improvement is more accentuated the more rapid and strong changes occur in the business
environment. Overall, these outcomes illustrate the high potential of this approach for
obtained some insights on the impact of the relevant variables on the suitability of the
inventory policies. In this regard, the best policy depends primarily on the cost structure of
the node. Moreover, our results reveal that the optimal policy is much more sensitive to the
inventory policy of the upper echelons than to that of the lower echelons of the supply
chain. Interestingly, we have noticed that the optimal policy of the wholesaler depends
heavily on whether, or not, the retailer’s policy mitigates the Bullwhip Effect.
different machine learning techniques applied to this problem. We plan analyze if the
additional complexity that other techniques entail (compared to inductive learning) derive
control techniques (Camacho and Bordons 2012) also defines a promising solution strategy
for the problem under consideration. Machine learning techniques could also be useful for
improving the control of inventories in contexts with inventory inaccuracies, i.e. deviations
between the actual and the recorded inventory (e.g. Li and Wang 2018). Another
interesting next step could be the exploration of the value of machine learning approaches
from the perspective of structural supply chain dynamics through the increasingly popular
concept of the ripple effect (see Dolgui, Ivanov, and Sokolov 2018). Finally, the adaptation
opportunities (e.g. Goltsos et al. 2018), may also be research directions worth pursuing.
24
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Notes
1. Net stock refers to the end-of-period on-hand inventory. Positive values represent excess
inventory (available to satisfy next period’s demand), while negative values represent
backlogs (unfulfilled demand that still needs to be met); see Disney and Lambrecht (2008).
2. This common cost model considers that a certain guaranteed capacity (GC) is available in
each period. If less than GC is needed, labour stands idle for a proportion of the period,
hence an opportunity cost is incurred. If more than GC is required, labour works overtime
3. The absolute gain criterion 𝑔1 , representing the information gained by a test 𝑇 with 𝑘
|𝐷 |
outcomes, is defined by 𝑔1 (𝐷, 𝑇) = 𝐻(𝐷) − ∑𝑘𝑖=1 |𝐷|𝑖 ∙ 𝐻(𝐷𝑖 ); and the relative gain
𝑔1
criterion 𝑔2 is defined by 𝑔2 (𝐷, 𝑇) = |𝐷𝑖 | |𝐷 |
; see Quinlan (1996). In this work,
𝑘
− ∑𝑖=1 |𝐷| ∙log2 ( |𝐷|𝑖 )
4. The roots of this work are in the models developed by Priore et al. (2001, 2003, 2006,
2010), which use different machine learning techniques for automatically modifying the
dispatching rules of flexible manufacturing systems over time. Shiue, Guh, and Lee (2012)
review similar approaches in the literature. These works show that this dynamic approach
is able to produce breakthrough improvements in performance over the same rules applied
statically. This encouraged us to adapt this approach to the supply chain field.
References
Akhbari, M., Y. Z. Mehrjerdi, H. K. Zare, and A. Makui. 2014. “A novel continuous KNN
prediction algorithm to improve manufacturing policies in a VMI supply chain.” International
Journal of Engineering, Transactions B: Applications 27 (11): 1681-1690.
25
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
APICS. 2011. APICS Operations Management Body of Knowledge Framework (Third Edition).
Chicago (IL): APICS The Association for Operations Management.
Axsäter, S. 2003. “Supply chain operations: Serial and distribution inventory systems.” Handbooks
in Operations Research and Management Science 11: 525-559.
Bajari, P., D. Nekipelov, S. P. Ryan, and M. Yang. 2015. “Machine Learning Methods for Demand
Estimation.” American Economic Review 105 (5): 481-85.
Basse, R. M., O. Charif, and K. Bódis. 2016. “Spatial and temporal dimensions of land use change
in cross border region of Luxembourg. Development of a hybrid approach integrating GIS, cellular
automata and decision learning tree models.” Applied Geography 67: 94-108.
Bischak, D. P., D. J. Robb, E. A. Silver, and J. D. Blackburn. 2014. “Analysis and Management of
Periodic Review, Order‐ Up‐ To Level Inventory Systems with Order Crossover.” Production and
Operations Management 23 (5): 762-772.
Camacho, E. F., and C. A. Bordons, C. A. 2012. Model predictive control in the process industry.
Berlin: Springer Science & Business Media.
Cannella, S., and E. Ciancimino. 2010. "On the bullwhip avoidance phase: supply chain
collaboration and order smoothing." International Journal of Production Research 48 (22): 6739-
6776.
Cannella, S., A. P. Barbosa-Póvoa, J. M. Framinan, and S. Relvas. 2013. “Metrics for bullwhip
effect analysis.” Journal of the Operational Research Society 64 (1): 1-16.
Cestnik, B., I. Kononenko, and I. Bratko. 1987. “ASSISTANT 86: A knowledge-elicitation tool for
sophisticated users.” In Progress in Machine Learning, edited by I. Bratko and N. Lavrac.
Wilmslow (UK): Sigma Press.
Chaharsooghi, S. K., J. Heydari, and S. H. Zegordi. 2008. “A reinforcement learning model for
supply chain ordering management: An application to the beer game.” Decision Support Systems
45: 949-959.
Chopra, S., and M. S. Sodhi. 2004. “Supply-chain breakdown.” MIT Sloan Management Review 46
(1): 53-61.
Ciancimino, E., S. Cannella, M. Bruccoleri, and J. M. Framinan. 2012. “On the bullwhip avoidance
phase: the synchronised supply chain.” European Journal of Operational Research 221 (1): 49-63.
26
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Disney, S. M., I. Farasyn, M. Lambrecht, D. R. Towill, and W. Van de Velde. 2006. “Taming the
bullwhip effect whilst watching customer service in a single supply chain echelon.” European
Journal of Operational Research 173 (1): 151-172.
Disney, S. M., G. Gaalman, and Hosoda. 2012. “Review of stochastic cost functions for production
and inventory control”. Paper presented at the 17th International Working Seminar of Production
Economics, Innsbruck, 117-128.
Disney, S. M., and M. R. Lambrecht. 2008. “On replenishment rules, forecasting and the bullwhip
effect in supply chains.” Foundations and Trends in Technology, Information, and Operations
Management 2 (1): 1–80.
Disney, S. M., A. Maltz, X. Wang, and R. D. Warburton. 2016. “Inventory management for
stochastic lead times with order crossovers.” European Journal of Operational Research 248 (2):
473-486.
Disney, S. M., and D. R. Towill. 2005. “Eliminating drift in inventory and order based production
control systems.” International Journal of Production Economics 93: 331-344.
Dolgui, A., D. Ivanov, and B. Sokolov. 2018. “Ripple effect in the supply chain: an analysis and
recent literature.” International Journal of Production Research 56(1-2), 414-430.
Filipič, B., and M. Junkar. 2000. “Using inductive machine learning to support decision making in
machining processes.” Computers in Industry 43 (1): 31-41.
Gaalman, G. 2006. “Bullwhip reduction for ARMA demand: The proportional order-up-to policy
versus the full-state-feedback policy.” Automatica 42 (8): 1283-1290.
Goltsos, T. E., B. Ponte, S. Wang, Y. Liu, M. M. Naim, and A. A. Syntetos. 2018. “The boomerang
returns? Accounting for the impact of uncertainties on the dynamics of remanufacturing systems.”
International Journal of Production Research, in press.
27
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Isaksson, O. H., and R. W. Seifert. 2016. “Quantifying the bullwhip effect using two-echelon data:
A cross-industry empirical investigation.” International Journal of Production Economics 171:
311-320.
Kahn, J. A. 1987. “Inventories and the volatility of production.” American Economic Review 77
(4): 667-679.
Khouja, M. 1999. “The single-period (news-vendor) problem: literature review and suggestions for
future research.” Omega 27 (5): 537-553.
Ko, M., A. Tiwari, and J. Mehnen. 2010. “A review of soft computing applications in supply chain
management.” Applied Soft Computing 10: 661-674.
Kuo, Y. H., and Kusiak, A. 2018. “From data to big data in production research: the past and future
trends.” International Journal of Production Research, in press.
Lancioni, R. A. 2000. “New developments in supply chain management for the millennium.”
Industrial Marketing Management 29 (1): 1-6.
Lee, H. L., V. Padmanabhan, and S. Whang. 1997. “The bullwhip effect in supply chains.” MIT
Sloan Management Review 38 (3): 93.
Li, Q., and S. M. Disney. 2017. “Revisiting rescheduling: MRP nervousness and the bullwhip
effect.” International Journal of Production Research 55 (7): 1992-2012.
Li, M., and Z. Wang. 2018. “An integrated robust replenishment/production/distribution policy
under inventory inaccuracy.” International Journal of Production Research 56 (12): 4115-4131.
Lim, T. S., W. Y. Loh, and Y. S. Shih. 2000. “A comparison of prediction accuracy, complexity,
and training time of thirty-three old and new classification algorithms.” Machine Learning 40 (3):
203-228.
Lin, J., M. M. Naim, L. Purvis, and J. Gosling, J. 2017. “The extension and exploitation of the
inventory and order-based production control system archetype from 1982 to 2015.” International
Journal of Production Economics 194: 135-152.
Lin, J., V. L. Spiegler, and M. M. Naim. 2018. “Dynamic analysis and design of a semiconductor
supply chain: a control engineering approach.” International Journal of Production Research 56
(13): 4585-4611.
28
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Melnyk, S. A., R. R. Lummus, R. J. Vokurka, L.J. Burns, and J. Sandor. 2009. "Mapping the future
of supply chain management: a Delphi study." International Journal of Production Research 47
(16): 4629-4653.
Metters, R. 1997. “Quantifying the bullwhip effect in supply chains.” Journal of Operations
Management 15 (2): 89-100.
Min, H. 2010. “Artificial intelligence in supply chain management: theory and applications.”
International Journal of Logistics Research and Applications 13 (1): 13-39.
Piramuthu, S. 2005b. “Machine learning for dynamic multi-product supply chain formation.”
Expert Systems with Applications 29 (4): 985-990.
Ponte, B., E. Sierra, D. de la Fuente, and J. Lozano. 2017. “Exploring the interaction of inventory
policies across the supply chain: An agent-based approach.” Computers and Operations Research
78: 335-348.
Ponte, B., X. Wang, D. de la Fuente, and S. M. Disney. 2017. “Exploring nonlinear supply chains:
the dynamics of capacity constraints.” International Journal of Production Research 55 (14): 4053-
4067.
Priore, P., D. de la Fuente, R. Pino, and J. Puente. 2001. “Learning-based scheduling of flexible
manufacturing systems using case-based reasoning.” Applied Artificial Intelligence 15: 949-963.
Priore, P., D. de la Fuente, R. Pino, and J. Puente. 2003. “Dynamic scheduling of flexible
manufacturing systems using neural networks and inductive learning.” Integrated Manufacturing
Systems 14 (2): 160-168.
29
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Priore, P., A. Gómez, R. Pino, and R. Rosillo. 2014. “Dynamic scheduling of manufacturing
systems using machine learning: an updated review.” Artificial Intelligence for Engineering
Design, Analysis and Manufacturing 28 (1): 83-97.
Priore, P., J. Parreño, R. Pino, A. Gómez, and J. Puente. 2010. “Learning-based scheduling of
flexible manufacturing systems using support vector machines.” Applied Artificial Intelligence 24:
194-209.
Puche, J., B. Ponte, J. Costas, R. Pino, and D. de la Fuente. 2016. “Systemic approach to supply
chain management through the viable system model and the theory of constraints.” Production
Planning & Control 27 (5): 421-430.
Quinlan, J. R. 1979. “Discovering rules by induction from large collections of examples”. In Expert
Systems in the Micro Electronic Age, edited by D. Michie. Edinburgh (UK): University Press.
Quinlan, J. R. 1993. C4.5: Programs for Machine Learning. San Mateo (CA): Morgan Kaufmann.
Quinlan, J. R. 1996. “Improved use of continuous attributes in C4. 5.” Journal of Artificial
Intelligence Research 4: 77-90.
Rendell, L. A. 1983. “A new basis for state-space learning systems and a successful
implementation.” Artificial Intelligence 20: 369-392.
Sillanpää, V., and J. Liesiö. 2018. “Forecasting replenishment orders in retail: value of modelling
low and intermittent consumer demand with distributions.” International Journal of Production
Research 56 (12): 4168-4185.
Steinker, S., M. Pesch, and K. Hoberg. 2016. “Inventory management under financial distress: an
empirical analysis.” International Journal of Production Research 54 (17): 5182-5207.
Shen, B., T. M. Choi, and S. Minner. 2018. “A review on supply chain contracting with
information considerations: information updating and information asymmetry.” International
Journal of Production Research, in press.
Shiue, Y. R., R. S. Guh, and K. C. Lee. 2012. “Development of machine learning‐ based real time
scheduling systems: using ensemble based on wrapper feature selection approach.” International
Journal of Production Research 50 (20): 5887-5905.
Sui, Z., A. Gosavi, and L. Lin. 2010. “A reinforcement learning approach for inventory
replenishment in vendor-managed inventory systems with consignment inventory.” Engineering
Management Journal 22 (4): 44-53.
30
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Syam, N., and A. Sharma. 2018. “Waiting for a sales renaissance in the fourth industrial revolution:
Machine learning and artificial intelligence in sales research and practice.”, Industrial Marketing
Management 69: 135-146.
Wang, X., and S. M. Disney. 2016. “The bullwhip effect: Progress, trends and directions.”
European Journal of Operational Research 250 (3): 691-701.
Witten, I. H., E. Frank, M. A. Hall, and C. J. Pal. 2016. Data Mining: Practical machine learning
tools and techniques. Cambridge (MA): Morgan Kaufmann.
Wu, X., V. Kumar, J. R. Quinlan, J. Ghosh, Q. Yang, H. Motoda, G. J. McLachlan, A. Ng, B. Liu,
S. Y. Philip, and Z. H. Zhou. 2008. “Top 10 algorithms in data mining.” Knowledge and
information systems 14 (1): 1-37.
Atributes Class
A1: A4:
Example A2: Tw A3: Tf A5: βr A6: βf A7: wo Policy
Tr CV
1 1 1 2 16% 0.1328 0.3434 0.1153 OUT
…
792 2 2 3 30% 0.8447 0.2830 0.4233 POUT_H
793 3 3 3 26% 0.1451 0.0091 0.1556 OUT
794 4 2 3 17% 0.6430 0.5161 0.7269 POUT_M
…
1466 3 1 2 34% 0.0356 0.2517 0.3727 OUT
1467 2 2 3 42% 0.8034 0.6466 0.8109 POUT_L
1468 2 3 4 23% 0.2413 0.3050 0.7704 POUT_H
…
2000 2 1 2 17% 0.4290 0.3362 0.9676 POUT_L
31
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
wo > 0.839 and βr > 0.063 and βr > 0.086 and wo >
1 POUT_L 263 / 298
0.842
Note: We emphasize in italics the best static policy. In parentheses, we show the improvement of the
32
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Note: We emphasize in italics the best static policy. In parentheses, we show the improvement of the
33
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
management
34
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
Figure 5. Relationship between the hit ratio and the number of examples in the training set.
35
“Applying Machine Learning to the Dynamic Selection of Replenishment Policies in Fast-Changing Supply Chain Environments”, by P. Priore, B..Ponte, R.
Rosillo and D. de la Fuente. Article accepted by the International Journal of Production Research. DOI: 10.1080/00207543.2018.1552369
36