Multidimensional assignment problem

Last updated

The multidimensional assignment problem (MAP) is a fundamental combinatorial optimization problem which was introduced by William Pierskalla. [1] This problem can be seen as a generalization of the linear assignment problem. [2] In words, the problem can be described as follows:

Contents

An instance of the problem has a number of agents (i.e., cardinality parameter) and a number of job characteristics (i.e., dimensionality parameter) such as task, machine, time interval, etc. For example, an agent can be assigned to perform task X, on machine Y, during time interval Z. Any agent can be assigned to perform a job with any combination of unique job characteristics at some cost. These costs may vary based on the assignment of agent to a combination of job characteristics - specific task, machine, time interval, etc. The problem is to minimize the total cost of assigning the agents so that the assignment of agents to each job characteristic is an injective function, or one-to-one function from agents to a given job characteristic.

Alternatively, describing the problem using graph theory:

The multidimensional assignment problem consists of finding, in a weighted multipartite graph, a matching of a given size, in which the sum of weights of the edges is minimum. [3]

Formal definition

Various formulations of this problem can be found in the literature. Using cost-functions, the dimensional assignment problem (or MAP) can be stated as follows:

Given sets, and , of equal size, together with a cost array or multidimensional weight function  : , find permutations  : A such that the total cost function:

is minimized. [4]

Problem parameters

The multidimensional assignment problem (MAP) has two key parameters that determine the size of a problem instance:

  1. The dimensionality parameter
  2. The cardinality parameter, where denotes the number of elements in .

Size of cost array

Any problem instance of the MAP with parameters has its specific cost array, which consists of instance-specific costs/weights parameters . is the size of cost array.

Number of feasible solutions

The feasible region or solution space of the MAP is very large. The number of feasible solutions (the size of the MAP instance) depends on the MAP parameters . Specifically, . [2]

Computational complexity

The problem is generally NP-hard. In other words, there is no known algorithm for solving this problem in polynomial time, and so a long computational time may be needed for solving problem instances of even moderate size (based on dimensionality and cardinality parameters). [5]

Applications

The problem found application in many domains:

Related Research Articles

<span class="mw-page-title-main">Knapsack problem</span> Problem in combinatorial optimization

The knapsack problem is the following problem in combinatorial optimization:

<span class="mw-page-title-main">Travelling salesman problem</span> NP-hard problem in combinatorial optimization

The travelling salesman problem, also known as the travelling salesperson problem (TSP), asks the following question: "Given a list of cities and the distances between each pair of cities, what is the shortest possible route that visits each city exactly once and returns to the origin city?" It is an NP-hard problem in combinatorial optimization, important in theoretical computer science and operations research.

In machine learning, support vector machines are supervised max-margin models with associated learning algorithms that analyze data for classification and regression analysis. Developed at AT&T Bell Laboratories by Vladimir Vapnik with colleagues SVMs are one of the most studied models, being based on statistical learning frameworks or VC theory proposed by Vapnik and Chervonenkis (1974).

<span class="mw-page-title-main">Assignment problem</span> Combinatorial optimization problem

The assignment problem is a fundamental combinatorial optimization problem. In its most general form, the problem is as follows:

An inverse problem in science is the process of calculating from a set of observations the causal factors that produced them: for example, calculating an image in X-ray computed tomography, source reconstruction in acoustics, or calculating the density of the Earth from measurements of its gravity field. It is called an inverse problem because it starts with the effects and then calculates the causes. It is the inverse of a forward problem, which starts with the causes and then calculates the effects.

<span class="mw-page-title-main">Multidimensional scaling</span> Set of related ordination techniques used in information visualization

Multidimensional scaling (MDS) is a means of visualizing the level of similarity of individual cases of a data set. MDS is used to translate distances between each pair of objects in a set into a configuration of points mapped into an abstract Cartesian space.

<span class="mw-page-title-main">Combinatorial optimization</span> Subfield of mathematical optimization

Combinatorial optimization is a subfield of mathematical optimization that consists of finding an optimal object from a finite set of objects, where the set of feasible solutions is discrete or can be reduced to a discrete set. Typical combinatorial optimization problems are the travelling salesman problem ("TSP"), the minimum spanning tree problem ("MST"), and the knapsack problem. In many such problems, such as the ones previously mentioned, exhaustive search is not tractable, and so specialized algorithms that quickly rule out large parts of the search space or approximation algorithms must be resorted to instead.

In computer science, parameterized complexity is a branch of computational complexity theory that focuses on classifying computational problems according to their inherent difficulty with respect to multiple parameters of the input or output. The complexity of a problem is then measured as a function of those parameters. This allows the classification of NP-hard problems on a finer scale than in the classical setting, where the complexity of a problem is only measured as a function of the number of bits in the input. This appears to have been first demonstrated in Gurevich, Stockmeyer & Vishkin (1984). The first systematic work on parameterized complexity was done by Downey & Fellows (1999).

The quadratic assignment problem (QAP) is one of the fundamental combinatorial optimization problems in the branch of optimization or operations research in mathematics, from the category of the facilities location problems first introduced by Koopmans and Beckmann.

A fully polynomial-time approximation scheme (FPTAS) is an algorithm for finding approximate solutions to function problems, especially optimization problems. An FPTAS takes as input an instance of the problem and a parameter ε > 0. It returns as output a value which is at least times the correct value, and at most times the correct value.

<span class="mw-page-title-main">Differential evolution</span> Method of mathematical optimization

In evolutionary computation, differential evolution (DE) is a method that optimizes a problem by iteratively trying to improve a candidate solution with regard to a given measure of quality. Such methods are commonly known as metaheuristics as they make few or no assumptions about the optimized problem and can search very large spaces of candidate solutions. However, metaheuristics such as DE do not guarantee an optimal solution is ever found.

Optimal job scheduling is a class of optimization problems related to scheduling. The inputs to such problems are a list of jobs and a list of machines. The required output is a schedule – an assignment of jobs to machines. The schedule should optimize a certain objective function. In the literature, problems of optimal job scheduling are often called machine scheduling, processor scheduling, multiprocessor scheduling, or just scheduling.

In applied mathematics, the maximum generalized assignment problem is a problem in combinatorial optimization. This problem is a generalization of the assignment problem in which both tasks and agents have a size. Moreover, the size of each task might vary from one agent to the other.

In the case of systems composed of subsystems, the classification of quantum-entangledstates is richer than in the bipartite case. Indeed, in multipartite entanglement apart from fully separable states and fully entangled states, there also exists the notion of partially separable states.

In the mathematical fields of graph theory and combinatorial optimization, the bipartite dimension or biclique cover number of a graph G = (VE) is the minimum number of bicliques (that is complete bipartite subgraphs), needed to cover all edges in E. A collection of bicliques covering all edges in G is called a biclique edge cover, or sometimes biclique cover. The bipartite dimension of G is often denoted by the symbol d(G).

<span class="mw-page-title-main">David Shmoys</span> American mathematician

David Bernard Shmoys is a Professor in the School of Operations Research and Information Engineering and the Department of Computer Science at Cornell University. He obtained his Ph.D. from the University of California, Berkeley in 1984. His major focus has been in the design and analysis of algorithms for discrete optimization problems.

In the study of algorithms, an LP-type problem is an optimization problem that shares certain properties with low-dimensional linear programs and that may be solved by similar algorithms. LP-type problems include many important optimization problems that are not themselves linear programs, such as the problem of finding the smallest circle containing a given set of planar points. They may be solved by a combination of randomized algorithms in an amount of time that is linear in the number of elements defining the problem, and subexponential in the dimension of the problem.

Biogeography-based optimization (BBO) is an evolutionary algorithm (EA) that optimizes a function by stochastically and iteratively improving candidate solutions with regard to a given measure of quality, or fitness function. BBO belongs to the class of metaheuristics since it includes many variations, and since it does not make any assumptions about the problem and can therefore be applied to a wide class of problems.

<span class="mw-page-title-main">Dispersive flies optimisation</span>

Dispersive flies optimisation (DFO) is a bare-bones swarm intelligence algorithm which is inspired by the swarming behaviour of flies hovering over food sources. DFO is a simple optimiser which works by iteratively trying to improve a candidate solution with regard to a numerical measure that is calculated by a fitness function. Each member of the population, a fly or an agent, holds a candidate solution whose suitability can be evaluated by their fitness value. Optimisation problems are often formulated as either minimisation or maximisation problems.

References

  1. 1 2 Pierskalla, William P. (1968). "Letter to the Editor—The Multidimensional Assignment Problem". Operations Research. 16 (2). INFORMS: 422–431. doi:10.1287/opre.16.2.422.
  2. 1 2 3 Kammerdiner, Alla; Semenov, Alexander; Pasiliao, Eduardo (2021). "Multidimensional Assignment Problem for multipartite entity resolution". arXiv: 2112.03346 [cs.DM].
  3. Natu, Shardul; Date, Ketan; Nagi, Rakesh (2020). "GPU-accelerated Lagrangian heuristic for multidimensional assignment problems with decomposable costs". Parallel Computing. 97: 102666. doi: 10.1016/j.parco.2020.102666 . ISSN   0167-8191. S2CID   221667518.
  4. Karapetyan, Daniel; Gutin, Gregory (2011-06-01). "Local search heuristics for the multidimensional assignment problem" (PDF). Journal of Heuristics. 17 (3): 201–249. doi:10.1007/s10732-010-9133-3. ISSN   1572-9397. S2CID   3446729.
  5. Nguyen, Duc Manh; Le Thi, Hoai An; Pham Dinh, Tao (2012-10-12). "Solving the Multidimensional Assignment Problem by a Cross-Entropy method". Journal of Combinatorial Optimization. 27 (4): 808–823. doi:10.1007/s10878-012-9554-z. ISSN   1382-6905. S2CID   254658376.
  6. Poore, Aubrey B. (1994). "Multidimensional assignment formulation of data association problems arising from multitarget and multisensor tracking". Computational Optimization and Applications. 3 (1): 27–57. doi: 10.1007/BF01299390 . S2CID   33848795.
  7. Pusztaszeri, Jean-François; Rensing, Paul E.; Liebling, Thomas M. (1996). "Tracking elementary particles near their primary vertex: a combinatorial approach". Journal of Global Optimization. 9 (1): 41–64. doi:10.1007/BF00121750. S2CID   2002168.
  8. Kammerdiner, Alla R.; Guererro, Andre N. (2019). "Data-driven combinatorial optimization for sensor-based assessment of near falls" . Annals of Operations Research. 276 (1–2): 137–153. doi:10.1007/s10479-017-2585-1. ISSN   0254-5330. S2CID   254223885.