Optimal computing budget allocation

Last updated

In Computer Science, Optimal Computing Budget Allocation (OCBA) is a simulation optimization method designed to maximize the Probability of Correct Selection (PCS) while minimizing computational costs. First introduced by Dr. Chun-Hung Chen in the mid-1990s, OCBA determines how many simulation runs (or how much computational time) or the number of replications each design alternative needs to identify the best option while using as few resources as possible. [1] [2] It works by focusing more on alternatives that are harder to evaluate, such as those with higher uncertainty or close performance to the best option.

Contents

Simply put, OCBA ensures that computational resources are distributed efficiently by allocating more simulation effort to design alternatives that are harder to evaluate or more likely to be the best. This allows researchers and decision-makers to achieve accurate results faster and with fewer resources.

OCBA has also been shown to enhance partition-based random search algorithms for solving deterministic global optimization problems. [3] Over the years, OCBA has been applied in manufacturing systems design, healthcare planning, and financial modeling. It has also been extended to handle more complex scenarios, such as balancing multiple objectives, [4] feasibility determination, [5] and constrained optimization. [6]

Intuitive Explanation

The goal of OCBA is to provide a systematic approach to efficiently run a large number of simulations by focusing only on the critical alternatives, in order to select the best alternative.

In other words, OCBA prioritizes only the most critical alternatives, minimizing computation time and reducing the variances of these critical estimators. The expected outcome is maintaining the required level of accuracy while requiring fewer computational resources. [7]

Figure 1: Preliminary simulation results show alternatives 2 and 3 have lower average delay times. OCBA suggests focusing further simulation resources on alternatives 2 and 3 while stopping simulations for alternatives 1, 4, and 5 to save costs without compromising accuracy. Comparing 5 different alternatives with respect to Cost.png
Figure 1: Preliminary simulation results show alternatives 2 and 3 have lower average delay times. OCBA suggests focusing further simulation resources on alternatives 2 and 3 while stopping simulations for alternatives 1, 4, and 5 to save costs without compromising accuracy.

For example, consider a simulation involving five alternatives, where the goal is to select the one with the minimum average delay time. Figure 1, shows preliminary simulation results (i.e., having run only a fraction of the required number of simulation replications). Alternatives 2 and 3 clearly have significantly lower delay times (highlighted in red). To save computation cost—which includes time, resources, and money spent on running simulations—OCBA suggests that more replications should be allocated to alternatives 2 and 3, while simulations for alternatives 1, 4, and 5 can be stopped much earlier without compromising accuracy.

Core Optimization Problem

Simulation is widely used for designing large, complex, stochastic systems, where analytical solutions are often infeasible. However, simulations can be computationally expensive because multiple simulation runs are needed to account for stochastic variability. The challenge lies in efficiently allocating limited computational resources to identify the best design alternative with high confidence.

The primary objective of OCBA is to maximize the Probability of Correct Selection (PCS), which represents the likelihood of identifying the best-performing design alternative among a finite set of options. This goal must be achieved while adhering to a limited computational budget. PCS is calculated based on the number of simulation replications allocated to each design.

The problem is mathematically formulated as:

Subject to:

where:

: Total number of design alternatives

: Number of simulation replications allocated to the -th design

: Total computational budget

OCBA optimizes the allocation of simulation replications by focusing on alternatives with higher variances or smaller performance gaps relative to the best alternative. The ratio of replications between two alternatives, such as and , is determined by the following formula:

Here:

: The variance of the performance of alternative .

: The performance gap between the best alternative () and alternative .

: The number of simulation replications allocated to alternative .

This formula ensures that alternatives with smaller performance gaps () or higher variances () receive more simulation replications. This maximizes computational efficiency while maintaining a high Probability of Correct Selection (PCS), ensuring computational efficiency by reducing replications for non-critical alternatives and increasing them for critical ones. [8] Numerical results show that OCBA can achieve the same simulation quality with only one-tenth of the computational effort compared to traditional methods. [2]

Some extensions of OCBA

Experts in the field explain that in some problems it is important to not only know the best alternative among a sample, but the top 5, 10, or even 50, because the decision maker may have other concerns that may affect the decision which are not modeled in the simulation.

According to Szechtman and Yücesan (2008), [9] OCBA is also helpful in feasibility determination problems. This is where the decisions makers are only interested in differentiating feasible alternatives from the infeasible ones. Further, choosing an alternative that is simpler, yet similar in performance is crucial for other decision makers. In this case, the best choice is among top-r simplest alternatives, whose performance rank above desired levels. [10]

In addition, Trailovic [11] and Pao [12] (2004) demonstrate an OCBA approach, where we find alternatives with minimum variance, instead of with best mean. Here, we assume unknown variances, voiding the OCBA rule (assuming that the variances are known). During 2010 research was done on an OCBA algorithm that is based on a t distribution. The results show no significant differences between those from t-distribution and normal distribution. The above presented extensions of OCBA is not a complete list and is yet to be fully explored and compiled. [2]

Multi-Objective OCBA

Multi-Objective Optimal Computing Budget Allocation (MOCBA) is the OCBA concept that applies to multi-objective problems. In a typical MOCBA, the PCS is defined as

in which

We notice that, the Type I error and Type II error for identifying a correct Pareto set are respectively

and .

Besides, it can be proven that

and

where is the number of objectives, and follows posterior distribution Noted that and are the average and standard deviation of the observed performance measures for objective of design , and is the number of observations.

Thus, instead of maximizing , we can maximize its lower bound, i.e., Assuming , the Lagrange method can be applied to conclude the following rules:

in which

and

Constrained optimization

Similar to the previous section, there are many situations with multiple performance measures. If the multiple performance measures are equally important, the decision makers can use the MOCBA. In other situations, the decision makers have one primary performance measure to be optimized while the secondary performance measures are constrained by certain limits.

The primary performance measure can be called the main objective while the secondary performance measures are referred as the constraint measures. This falls into the problem of constrained optimization. When the number of alternatives is fixed, the problem is called constrained ranking and selection where the goal is to select the best feasible design given that both the main objective and the constraint measures need to be estimated via stochastic simulation. The OCBA method for constrained optimization (called OCBA-CO) can be found in Pujowidianto et al. (2009) [13] and Lee et al. (2012). [14]

The key change is in the definition of PCS. There are two components in constrained optimisation, namely optimality and feasibility. As a result, the simulation budget can be allocated to each non-best design either based on the optimality or feasibility. In other word, a non-best design will not be wrongly selected as the best feasible design if it remains either infeasible or worse than the true best feasible design. The idea is that it is not necessary to spend a large portion of the budget to determine the feasibility if the design is clearly worse than the best. Similarly, we can save the budget by allocating based on feasibility if the design is already better than the best in terms of the main objective.

Feasibility determination

The goal of this problem is to determine all the feasible designs from a finite set of design alternatives, where the feasible designs are defined as the designs with their performance measures satisfying specified control requirements (constraints). With all the feasible designs selected, the decision maker can easily make the final decision by incorporating other performance considerations (e.g., deterministic criteria, such as cost, or qualitative criteria which are difficult to mathematically evaluate). Although the feasibility determination problem involves stochastic constraints too, it is distinguished from the constrained optimization problems introduced above, in that it aims to identify all the feasible designs instead of the single best feasible one.

Define

Suppose all the constraints are provided in form of , . The probability of correctly selecting all the feasible designs is

and the budget allocation problem for feasibility determination is given by Gao and Chen (2017) [15]

Let and . The asymptotic optimal budget allocation rule is

Intuitively speaking, the above allocation rule says that (1) for a feasible design, the dominant constraint is the most difficult one to be correctly detected among all the constraints; and (2) for an infeasible design, the dominant constraint is the easiest one to be correctly detected among all constraints.

OCBA with expected opportunity cost

The original OCBA maximizes the probability of correct selection (PCS) of the best design. In practice, another important measure is the expected opportunity cost (EOC), which quantifies how far away the mean of the selected design is from that of the real best. This measure is important because optimizing EOC not only maximizes the chance of selecting the best design but also ensures that the mean of the selected design is not too far from that of the best design, if it fails to find the best one. Compared to PCS, EOC penalizes a particularly bad choice more than a slightly incorrect selection, and is thus preferred by risk-neutral practitioners and decision makers.

Specifically, the expected opportunity cost is

where,

The budget allocation problem with the EOC objective measure is given by Gao et al. (2017) [16]

where is the proportion of the total simulation budget allocated to design . If we assume for all , the asymptotic optimal budget allocation rule for this problem is

where is the variance of the simulation samples of design . This allocation rule is the same as the asymptotic optimal solution of problem (1). That is, asymptotically speaking, maximizing PCS and minimizing EOC are the same thing.

OCBA with input uncertainty

An implicit assumption for the aforementioned OCBA methods is that the true input distributions and their parameters are known, while in practice, they are typically unknown and have to be estimated from limited historical data. This may lead to uncertainty in the estimated input distributions and their parameters, which might (severely) affect the quality of the selection. Assuming that the uncertainty set contains a finite number of scenarios for the underlying input distributions and parameters, Gao et al. (2017) [17] introduces a new OCBA approach by maximizing the probability of correctly selecting the best design under a fixed simulation budget, where the performance of a design is measured by its worst-case performance among all the possible scenarios in the uncertainty set.

Recent Applications of OCBA

Optimal Computing Budget Allocation (OCBA), has continued to evolve, demonstrating its adaptability and efficiency in addressing complex decision-making problems across various domains.

These recent innovations demonstrate OCBA's growing versatility and effectiveness in optimizing resource allocation for diverse applications.

Emerging Research Area: Integration of Machine Learning with OCBA

The integration of Machine Learning (ML) with Optimal Computing Budget Allocation (OCBA) represents a promising area of research, leveraging ML’s predictive capabilities to enhance the efficiency and accuracy of simulation optimization. By incorporating ML models, OCBA can dynamically adapt resource allocation strategies, addressing complex decision-making problems with greater computational efficiency.

Applications

Predictive Multi-Fidelity Models: Gaussian Mixture Models (GMMs) predict relationships between low- and high-fidelity simulations, enabling OCBA to focus on the most promising alternatives. Multi-fidelity models combine insights from low-fidelity simulations, which are computationally inexpensive but less accurate, and high-fidelity simulations, which are more accurate but computationally intensive. The integration of GMMs into this process allows OCBA to strategically allocate computational resources across fidelity levels, significantly reducing simulation costs while maintaining decision accuracy. [21]

Dynamic Resource Allocation in Healthcare: A Bayesian OCBA framework has been applied to allocate resources in hospital emergency departments, balancing service quality with operational efficiency. By minimizing expected opportunity costs, this approach supports real-time decision-making in high-stakes environments. [22] Additionally, the integration of OCBA with real-time digital twin-based optimization has further advanced its application in predictive simulation learning, enabling dynamic adjustments to resource allocation in healthcare settings. [23] Furthermore, a contextual ranking and selection method for personalized medicine leverages OCBA to optimize resource allocation in treatments tailored to individual patient profiles, demonstrating its potential in personalized healthcare. [24]

Related Research Articles

<span class="mw-page-title-main">Pauli matrices</span> Matrices important in quantum mechanics and the study of spin

In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.

<span class="mw-page-title-main">Pareto distribution</span> Probability distribution

The Pareto distribution, named after the Italian civil engineer, economist, and sociologist Vilfredo Pareto, is a power-law probability distribution that is used in description of social, quality control, scientific, geophysical, actuarial, and many other types of observable phenomena; the principle originally applied to describing the distribution of wealth in a society, fitting the trend that a large portion of wealth is held by a small fraction of the population. The Pareto principle or "80-20 rule" stating that 80% of outcomes are due to 20% of causes was named in honour of Pareto, but the concepts are distinct, and only Pareto distributions with shape value of log45 ≈ 1.16 precisely reflect it. Empirical observation has shown that this 80-20 distribution fits a wide range of cases, including natural phenomena and human activities.

In the field of mathematical optimization, stochastic programming is a framework for modeling optimization problems that involve uncertainty. A stochastic program is an optimization problem in which some or all problem parameters are uncertain, but follow known probability distributions. This framework contrasts with deterministic optimization, in which all problem parameters are assumed to be known exactly. The goal of stochastic programming is to find a decision which both optimizes some criteria chosen by the decision maker, and appropriately accounts for the uncertainty of the problem parameters. Because many real-world decisions involve uncertainty, stochastic programming has found applications in a broad range of areas ranging from finance to transportation to energy optimization.

In physics, the Hamilton–Jacobi equation, named after William Rowan Hamilton and Carl Gustav Jacob Jacobi, is an alternative formulation of classical mechanics, equivalent to other formulations such as Newton's laws of motion, Lagrangian mechanics and Hamiltonian mechanics.

Mehrotra's predictor–corrector method in optimization is a specific interior point method for linear programming. It was proposed in 1989 by Sanjay Mehrotra.

Fairness measures or metrics are used in network engineering to determine whether users or applications are receiving a fair share of system resources. There are several mathematical and conceptual definitions of fairness.

In mathematical finance, the SABR model is a stochastic volatility model, which attempts to capture the volatility smile in derivatives markets. The name stands for "stochastic alpha, beta, rho", referring to the parameters of the model. The SABR model is widely used by practitioners in the financial industry, especially in the interest rate derivative markets. It was developed by Patrick S. Hagan, Deep Kumar, Andrew Lesniewski, and Diana Woodward.

Covariance matrix adaptation evolution strategy (CMA-ES) is a particular kind of strategy for numerical optimization. Evolution strategies (ES) are stochastic, derivative-free methods for numerical optimization of non-linear or non-convex continuous optimization problems. They belong to the class of evolutionary algorithms and evolutionary computation. An evolutionary algorithm is broadly based on the principle of biological evolution, namely the repeated interplay of variation and selection: in each generation (iteration) new individuals are generated by variation of the current parental individuals, usually in a stochastic way. Then, some individuals are selected to become the parents in the next generation based on their fitness or objective function value . Like this, individuals with better and better -values are generated over the generation sequence.

Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.

The Gittins index is a measure of the reward that can be achieved through a given stochastic process with certain properties, namely: the process has an ultimate termination state and evolves with an option, at each intermediate state, of terminating. Upon terminating at a given state, the reward achieved is the sum of the probabilistic expected rewards associated with every state from the actual terminating state to the ultimate terminal state, inclusive. The index is a real scalar.

In computational chemistry, a constraint algorithm is a method for satisfying the Newtonian motion of a rigid body which consists of mass points. A restraint algorithm is used to ensure that the distance between mass points is maintained. The general steps involved are: (i) choose novel unconstrained coordinates, (ii) introduce explicit constraint forces, (iii) minimize constraint forces implicitly by the technique of Lagrange multipliers or projection methods.

<span class="mw-page-title-main">Truncated normal distribution</span> Type of probability distribution

In probability and statistics, the truncated normal distribution is the probability distribution derived from that of a normally distributed random variable by bounding the random variable from either below or above. The truncated normal distribution has wide applications in statistics and econometrics.

<span class="mw-page-title-main">Logit-normal distribution</span> Probability distribution

In probability theory, a logit-normal distribution is a probability distribution of a random variable whose logit has a normal distribution. If Y is a random variable with a normal distribution, and t is the standard logistic function, then X = t(Y) has a logit-normal distribution; likewise, if X is logit-normally distributed, then Y = logit(X)= log (X/(1-X)) is normally distributed. It is also known as the logistic normal distribution, which often refers to a multinomial logit version (e.g.).

<span class="mw-page-title-main">Reinforced solid</span>

In solid mechanics, a reinforced solid is a brittle material that is reinforced by ductile bars or fibres. A common application is reinforced concrete. When the concrete cracks the tensile force in a crack is not carried any more by the concrete but by the steel reinforcing bars only. The reinforced concrete will continue to carry the load provided that sufficient reinforcement is present. A typical design problem is to find the smallest amount of reinforcement that can carry the stresses on a small cube. This can be formulated as an optimization problem.

In computational chemistry and molecular dynamics, the combination rules or combining rules are equations that provide the interaction energy between two dissimilar non-bonded atoms, usually for the part of the potential representing the van der Waals interaction. In the simulation of mixtures, the choice of combining rules can sometimes affect the outcome of the simulation.

Quantum optimization algorithms are quantum algorithms that are used to solve optimization problems. Mathematical optimization deals with finding the best solution to a problem from a set of possible solutions. Mostly, the optimization problem is formulated as a minimization problem, where one tries to minimize an error which depends on the solution: the optimal solution has the minimal error. Different optimization techniques are applied in various fields such as mechanics, economics and engineering, and as the complexity and amount of data involved rise, more efficient ways of solving optimization problems are needed. Quantum computing may allow problems which are not practically feasible on classical computers to be solved, or suggest a considerable speed up with respect to the best known classical algorithm.

Batch normalization is a method used to make training of artificial neural networks faster and more stable through normalization of the layers' inputs by re-centering and re-scaling. It was proposed by Sergey Ioffe and Christian Szegedy in 2015.

An additive process, in probability theory, is a cadlag, continuous in probability stochastic process with independent increments. An additive process is the generalization of a Lévy process. An example of an additive process that is not a Lévy process is a Brownian motion with a time-dependent drift. The additive process was introduced by Paul Lévy in 1937.

<span class="mw-page-title-main">Consensus based optimization</span> Iterative simulation method

Consensus-based optimization (CBO) is a multi-agent derivative-free optimization method, designed to obtain solutions for global optimization problems of the form

<span class="mw-page-title-main">Deep backward stochastic differential equation method</span>

Deep backward stochastic differential equation method is a numerical method that combines deep learning with Backward stochastic differential equation (BSDE). This method is particularly useful for solving high-dimensional problems in financial derivatives pricing and risk management. By leveraging the powerful function approximation capabilities of deep neural networks, deep BSDE addresses the computational challenges faced by traditional numerical methods in high-dimensional settings.

References

  1. Chen, Chun-Hung (1995). "An Effective Approach to Smartly Allocate Computing Budget for Discrete Event Simulation". Proceedings of the 34th IEEE Conference on Decision and Control. IEEE. pp. 2598–2605.
  2. 1 2 3 4 Chen, Chun-Hung; Lee, Loo H. (2011). Stochastic Simulation Optimization: An Optimal Computing Budget Allocation. World Scientific Series on Nonlinear Science Series A. Vol. 82. World Scientific. doi:10.1142/7437. ISBN   978-981-4282-64-2.
  3. Chen, Wei; Gao, Siyang; Chen, Chun-Hung; Shi, Lei (2014). "An Optimal Sample Allocation Strategy for Partition-Based Random Search". IEEE Transactions on Automation Science and Engineering. 11 (1). IEEE: 177–186. doi:10.1109/TASE.2013.2251881.
  4. Lee, Loo Hay; Li, Li Wei; Chen, Chun-Hung; Yap, C. M. (2012). "Approximation Simulation Budget Allocation for Selecting the Best Design in the Presence of Stochastic Constraints". IEEE Transactions on Automatic Control. 57 (12): 2940–2945. doi:10.1109/TAC.2012.2204478 (inactive 15 December 2024).{{cite journal}}: CS1 maint: DOI inactive as of December 2024 (link)
  5. Szechtman, R.; Yücesan, E. (2008). "A New Perspective on Feasibility Determination" (PDF). Proceedings of the 2008 Winter Simulation Conference. pp. 273–280.
  6. Gao, Shu; Xiao, Hongsheng; Zhou, Enlu; Chen, Wei (2017). "Robust Ranking and Selection with Optimal Computing Budget Allocation". Automatica. 81: 30–36. doi:10.1016/j.automatica.2017.03.015.
  7. Chen, Chun-Hung. "Optimal Computing Budget Allocation (OCBA) for Simulation-based Decision Making Under Uncertainty". Archived from the original on 1 October 2013. Retrieved 9 July 2013.
  8. Chen, C. H., J. Lin, E. Yücesan, and S. E. Chick, "Simulation Budget Allocation for Further Enhancing the Efficiency of Ordinal Optimization," Journal of Discrete Event Dynamic Systems, 2000.
  9. Szechtman R, Yücesan E (2008) A new perspective on feasibility determination. Proc of the 2008 Winter Simul Conf 273–280
  10. Jia QS, Zhou E, Chen CH (2012). efficient computing budget allocation for finding simplest good designs. IIE Trans, To Appear.
  11. Trailovic Tekin E, Sabuncuoglu I (2004) Simulation optimization: A comprehensive review on theory and applications. IIE Trans 36:1067–1081
  12. Trailovic L, Pao LY (2004) Computing budget allocation for efficient ranking and selection of variances with application to target tracking algorithms, IEEE Trans Autom Control 49:58–67.
  13. Pujowidianto NA, Lee LH, Chen CH, Yap CM (2009) Optimal computing budget allocation for constrained optimization. Proc of the 2009 Winter Simul Conf 584–589.
  14. Lee LH, Pujowidianto NA, Li LW, Chen CH, Yap CM (2012) Approximation simulation budget allocation for selecting the best design in the presence of stochastic constraints, IEEE Trans Autom Control 57:2940–2945.
  15. Gao, S. and W. Chen, "Efficient feasibility determination with multiple performance measure constraints," IEEE Transactions on Automatic Control, 62, 113–122, 2017.
  16. Gao, Siyang; Chen, Weiwei; Shi, Leyuan (2017). "A New Budget Allocation Framework for the Expected Opportunity Cost". Operations Research. 65 (3): 787–803. doi:10.1287/opre.2016.1581.
  17. Gao, S., H. Xiao, E. Zhou and W. Chen, "Robust Ranking and Selection with Optimal Computing Budget Allocation," Automatica, 81, 30–36, 2017.
  18. Chen, Chun-Hung (1995). "An Effective Approach to Smartly Allocate Computing Budget for Discrete Event Simulation". Proceedings of the 34th IEEE Conference on Decision and Control. IEEE. pp. 2598–2605.
  19. Li, Yunchuan; Fu, Michael C.; Xu, Jie (2020). "An Optimal Computing Budget Allocation Tree Policy for Monte Carlo Tree Search". arXiv preprint. 2009 (12407). arXiv: 2009.12407 .
  20. Wang, Yu; Tang, Wei; Yao, Yan; Zhu, Fang (2019). "DA-OCBA: Distributed Asynchronous Optimal Computing Budget Allocation Algorithm of Simulation Optimization Using Cloud Computing". Symmetry. 11 (10): 1297. Bibcode:2019Symm...11.1297W. doi: 10.3390/sym11101297 .
  21. Peng, Y.; Xu, J.; Lee, L. H.; Hu, J.; Chen, C. H. (2019). "Efficient Simulation Sampling Allocation Using Multifidelity Models". IEEE Transactions on Automatic Control. 64 (8): 3156–3169. doi:10.1109/TAC.2018.2886165.
  22. Chen, Weiwei; Gao, Siyang; Chen, Wenjie; Du, Jianzhong (2023). "Optimizing Resource Allocation in Service Systems via Simulation: A Bayesian Formulation". Production and Operations Management. 32: 65–81. doi:10.1111/poms.13825.
  23. Goodwin, Timothy; Xu, Jie; Celik, Niyazi; Chen, Chun-Hung (2024). "Real-Time Digital Twin-Based Optimization with Predictive Simulation Learning". Journal of Simulation. 18 (1): 47–64. doi:10.1080/17477778.2022.2046520.
  24. Gao, Siyang; Du, Jianzhong; Chen, Chun-Hung (2024). "A Contextual Ranking and Selection Method for Personalized Medicine". Manufacturing and Service Operations Management. 26 (1): 167–181. arXiv: 2206.12640 . doi:10.1287/msom.2022.0232.