Economic lot scheduling problem

Last updated

The economic lot scheduling problem (ELSP) is a problem in operations management and inventory theory that has been studied by many researchers for more than 50 years. The term was first used in 1958 by professor Jack D. Rogers of Berkeley, [1] who extended the economic order quantity model to the case where there are several products to be produced on the same machine, so that one must decide both the lot size for each product and when each lot should be produced. The method illustrated by Jack D. Rogers draws on a 1956 paper from Welch, W. Evert. [2] The ELSP is a mathematical model of a common issue for almost any company or industry: planning what to manufacture, when to manufacture and how much to manufacture.

Contents

Model formulation

The classic ELSP is concerned with scheduling the production of several products on a single machine in order to minimize the total costs incurred (which include setup costs and inventory holding costs).

We assume a known, non-varying demand for the m products (for example, there might be m=3 products and customers require 7 items a day of Product 1, 5 items a day of Product 2 and 2 items a day of Product 3). Customer demand is met from inventory and the inventory is replenished by our production facility.

A single machine is available which can make all the products, but not in a perfectly interchangeable way. Instead the machine needs to be set up to produce one product, incurring a setup cost and/or setup time, after which it will produce this product at a known rate . When it is desired to produce a different product, the machine is stopped and another costly setup is required to begin producing the next product. Let be the setup cost when switching from product i to product j and inventory cost is charged based on average inventory level of each item. N is the number of runs made, U the use rate, L the lot size and T the planning period.

To give a very concrete example, the machine might be a bottling machine and the products could be cases of bottled apple juice, orange juice and milk. The setup corresponds to the process of stopping the machine, cleaning it out and loading the tank of the machine with the desired fluid. This product switching must not be done too often or the setup costs will be large, but equally too long a production run of apple juice would be undesirable because it would lead to a large inventory investment and carrying cost for unsold cases of apple juice and perhaps stock-outs in orange juice and milk. The ELSP seeks the optimal trade off between these two extremes.

Rogers algorithm

1.Define:

= use period
cL=, the unit cost for a lot of size L
the total cost for N lots. To obtain the optimum:
Which yields as the optimum lot size. Now let:
be the total cost for NL±alots of size L±a
be the incremental cost of changing from size L to L+a
be the incremental cost of changing from size L to L-a

2.

Total quantity of an item required = UT
Total production time for an item = UT/P
Check that productive capacity is satisfied:

3.Compute:

as a whole number
If for a certain item, θ0 is not an even number, calculate:
And change L0 to L in the direction which incurs the least cost increase between +Δ and -Δ

4.Compute tp=L/P for each item and list items in order of increasing θ=L/U

5.For each pair of items ij check:

To forms pairs take the ith with the i+1th, i+2th, etc. If any of these inequalities is violated, calculate +Δ and -Δ for lot size increments of 2U and in order of size of cost change make step-by-step lot size changes. Repeat this step until both inequalities are satisfied.

6.

  1. Form all possible pairs as in Step 5
  2. For each pair, select θi < θj
  3. Determine whether tpi > tpj, tpi < tpj or tpi = tpj
  4. Select a value for eij(eij=0,1,2,3,...,θi - tpi - tpj) and calculate tpi+e and tpj+e
  5. Calculate Miθi-Mjθj by setting Mi=k and Mj=1,2,3,...,T/θj; ∀k∈(1,2,...,T/θi). Then check if one of the following boundary conditions is satisfied:
for or
for
If none of the boundary conditions is satisfied then eij is non-interfering: if i=1 in eij, pick the next larger e in sub-step 4, if i≠1 go back to sub-step 2. If some boundary condition is satisfied go to sub-step 4. If, for any pair, no non-interfering e appears, go back to Step 5.

7.Enter items in schedule and check it's feasibility

Stochastic ELSP

Of great importance in practice is to design, plan and operate shared capacity across multiple products with changeover times and costs in an uncertain demand environment. Beyond the selection of (expected) cycle times, with some amount of slack designed in ("safety time"), one has to also consider the amount of safety stock (buffer stock) that is needed to meet desired service level. [3]

Problem status

The problem is well known in the operations research community, and a large body of academic research work has been created to improve the model and to create new variations that solve specific issues.

The model is known as a NP-hard problem since it is not currently possible to find the optimal solution without checking nearly every possibility. What has been done follows two approaches: restricting the solution to be of a specific type (which makes it possible to find the optimal solution for the narrower problem), or approximate solution of the full problem using heuristics or genetic algorithms. [4]

See also

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

In complex analysis, a branch of mathematics, analytic continuation is a technique to extend the domain of definition of a given analytic function. Analytic continuation often succeeds in defining further values of a function, for example in a new region where the infinite series representation which initially defined the function becomes divergent.

<span class="mw-page-title-main">Gamma distribution</span> Probability distribution

In probability theory and statistics, the gamma distribution is a two-parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-squared distribution are special cases of the gamma distribution. There are two equivalent parameterizations in common use:

  1. With a shape parameter k and a scale parameter θ
  2. With a shape parameter and an inverse scale parameter , called a rate parameter.

In numerical analysis and computational statistics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the acceptance-rejection method or "accept-reject algorithm" and is a type of exact simulation method. The method works for any distribution in with a density.

<span class="mw-page-title-main">Cramér–Rao bound</span> Lower bound on variance of an estimator

In estimation theory and statistics, the Cramér–Rao bound (CRB) relates to estimation of a deterministic parameter. The result is named in honor of Harald Cramér and C. R. Rao, but has also been derived independently by Maurice Fréchet, Georges Darmois, and by Alexander Aitken and Harold Silverstone. It states that the precision of any unbiased estimator is at most the Fisher information; or (equivalently) the reciprocal of the Fisher information is a lower bound on its variance.

<span class="mw-page-title-main">Mechanism design</span> Field in game theory

Mechanism design is a field in economics and game theory that takes an objectives-first approach to designing economic mechanisms or incentives, toward desired objectives, in strategic settings, where players act rationally. Because it starts at the end of the game, then goes backwards, it is also called reverse game theory. It has broad applications, from economics and politics in fields such as market design, auction theory and social choice theory to networked-systems.

In physics, the Hamilton–Jacobi equation, named after William Rowan Hamilton and Carl Gustav Jacob Jacobi, is an alternative formulation of classical mechanics, equivalent to other formulations such as Newton's laws of motion, Lagrangian mechanics and Hamiltonian mechanics.

In statistics, G-tests are likelihood-ratio or maximum likelihood statistical significance tests that are increasingly being used in situations where chi-squared tests were previously recommended.

In mathematics, the Riesz–Thorin theorem, often referred to as the Riesz–Thorin interpolation theorem or the Riesz–Thorin convexity theorem, is a result about interpolation of operators. It is named after Marcel Riesz and his student G. Olof Thorin.

In statistics, the score test assesses constraints on statistical parameters based on the gradient of the likelihood function—known as the score—evaluated at the hypothesized parameter value under the null hypothesis. Intuitively, if the restricted estimator is near the maximum of the likelihood function, the score should not differ from zero by more than sampling error. While the finite sample distributions of score tests are generally unknown, they have an asymptotic χ2-distribution under the null hypothesis as first proved by C. R. Rao in 1948, a fact that can be used to determine statistical significance.

In mathematics, differential algebra is, broadly speaking, the area of mathematics consisting in the study of differential equations and differential operators as algebraic objects in view of deriving properties of differential equations and operators without computing the solutions, similarly as polynomial algebras are used for the study of algebraic varieties, which are solution sets of systems of polynomial equations. Weyl algebras and Lie algebras may be considered as belonging to differential algebra.

In mathematics, the Weyl character formula in representation theory describes the characters of irreducible representations of compact Lie groups in terms of their highest weights. It was proved by Hermann Weyl. There is a closely related formula for the character of an irreducible representation of a semisimple Lie algebra. In Weyl's approach to the representation theory of connected compact Lie groups, the proof of the character formula is a key step in proving that every dominant integral element actually arises as the highest weight of some irreducible representation. Important consequences of the character formula are the Weyl dimension formula and the Kostant multiplicity formula.

Proportional hazards models are a class of survival models in statistics. Survival models relate the time that passes, before some event occurs, to one or more covariates that may be associated with that quantity of time. In a proportional hazards model, the unique effect of a unit increase in a covariate is multiplicative with respect to the hazard rate. For example, taking a drug may halve one's hazard rate for a stroke occurring, or, changing the material from which a manufactured component is constructed may double its hazard rate for failure. Other types of survival models such as accelerated failure time models do not exhibit proportional hazards. The accelerated failure time model describes a situation where the biological or mechanical life history of an event is accelerated.

The sequential probability ratio test (SPRT) is a specific sequential hypothesis test, developed by Abraham Wald and later proven to be optimal by Wald and Jacob Wolfowitz. Neyman and Pearson's 1933 result inspired Wald to reformulate it as a sequential analysis problem. The Neyman-Pearson lemma, by contrast, offers a rule of thumb for when all the data is collected.

Stochastic approximation methods are a family of iterative methods typically used for root-finding problems or for optimization problems. The recursive update rules of stochastic approximation methods can be used, among other things, for solving linear systems when the collected data is corrupted by noise, or for approximating extreme values of functions which cannot be computed directly, but only estimated via noisy observations.

A ratio distribution is a probability distribution constructed as the distribution of the ratio of random variables having two other known distributions. Given two random variables X and Y, the distribution of the random variable Z that is formed as the ratio Z = X/Y is a ratio distribution.

<span class="mw-page-title-main">Kicked rotator</span>

The kicked rotator, also spelled as kicked rotor, is a paradigmatic model for both Hamiltonian chaos and quantum chaos. It describes a free rotating stick in an inhomogeneous "gravitation like" field that is periodically switched on in short pulses. The model is described by the Hamiltonian

Derrick's theorem is an argument by physicist G. H. Derrick which shows that stationary localized solutions to a nonlinear wave equation or nonlinear Klein–Gordon equation in spatial dimensions three and higher are unstable.

In quantum information theory, the Wehrl entropy, named after Alfred Wehrl, is a classical entropy of a quantum-mechanical density matrix. It is a type of quasi-entropy defined for the Husimi Q representation of the phase-space quasiprobability distribution. See for a comprehensive review of basic properties of classical, quantum and Wehrl entropies, and their implications in statistical mechanics.

Pure inductive logic (PIL) is the area of mathematical logic concerned with the philosophical and mathematical foundations of probabilistic inductive reasoning. It combines classical predicate logic and probability theory. Probability values are assigned to sentences of a first-order relational language to represent degrees of belief that should be held by a rational agent. Conditional probability values represent degrees of belief based on the assumption of some received evidence.

References

  1. Jack D. Rogers: A Computational Approach to the Economic Lot Scheduling Problem, Management Science, Vol. 4, No. 3, April 1958, pp. 264–291
  2. Welch, W. Evert, A Case of Simple Linear Programming, Management Methods 1956 in Jack D. Rogers: A Computational Approach to the Economic Lot Scheduling Problem, Management Science, Vol. 4, No. 3, April 1958, pp. 264–291
  3. Tayur, S. (2000). "Improving Operations and Quoting Accurate Lead Times in a Laminate Plant". Interfaces. 30 (5): 1–15. doi:10.1287/inte.30.5.1.11637.
  4. Zipkin Paul H., Foundations of Inventory Management, Boston: McGraw Hill, 2000, ISBN   0-256-11379-3

Further reading