Submodular set function

Last updated

In mathematics, a submodular set function (also known as a submodular function) is a set function that, informally, describes the relationship between a set of inputs and an output, where adding more of one input has a decreasing additional benefit (diminishing returns). The natural diminishing returns property which makes them suitable for many applications, including approximation algorithms, game theory (as functions modeling user preferences) and electrical networks. Recently, submodular functions have also found utility in several real world problems in machine learning and artificial intelligence, including automatic summarization, multi-document summarization, feature selection, active learning, sensor placement, image collection summarization and many other domains. [1] [2] [3] [4]

Contents

Definition

If is a finite set, a submodular function is a set function , where denotes the power set of , which satisfies one of the following equivalent conditions. [5]

  1. For every with and every we have that .
  2. For every we have that .
  3. For every and such that we have that .

A nonnegative submodular function is also a subadditive function, but a subadditive function need not be submodular. If is not assumed finite, then the above conditions are not equivalent. In particular a function defined by if is finite and if is infinite satisfies the first condition above, but the second condition fails when and are infinite sets with finite intersection.

Types and examples of submodular functions

Monotone

A set function is monotone if for every we have that . Examples of monotone submodular functions include:

Linear (Modular) functions
Any function of the form is called a linear function. Additionally if then f is monotone.
Budget-additive functions
Any function of the form for each and is called budget additive. [6]
Coverage functions
Let be a collection of subsets of some ground set . The function for is called a coverage function. This can be generalized by adding non-negative weights to the elements.
Entropy
Let be a set of random variables. Then for any we have that is a submodular function, where is the entropy of the set of random variables , a fact known as Shannon's inequality. [7] Further inequalities for the entropy function are known to hold, see entropic vector.
Matroid rank functions
Let be the ground set on which a matroid is defined. Then the rank function of the matroid is a submodular function. [8]

Non-monotone

A submodular function that is not monotone is called non-monotone.

Symmetric

A non-monotone submodular function is called symmetric if for every we have that . Examples of symmetric non-monotone submodular functions include:

Graph cuts
Let be the vertices of a graph. For any set of vertices let denote the number of edges such that and . This can be generalized by adding non-negative weights to the edges.
Mutual information
Let be a set of random variables. Then for any we have that is a submodular function, where is the mutual information.

Asymmetric

A non-monotone submodular function which is not symmetric is called asymmetric.

Directed cuts
Let be the vertices of a directed graph. For any set of vertices let denote the number of edges such that and . This can be generalized by adding non-negative weights to the directed edges.

Continuous extensions of submodular set functions

Often, given a submodular set function that describes the values of various sets, we need to compute the values of fractional sets. For example: we know that the value of receiving house A and house B is V, and we want to know the value of receiving 40% of house A and 60% of house B. To this end, we need a continuous extension of the submodular set function.

Formally, a set function with can be represented as a function on , by associating each with a binary vector such that when , and otherwise. A continuous extension of is a continuous function , that matches the value of on , i.e. .

Several kinds of continuous extensions of submodular functions are commonly used, which are described below.

Lovász extension

This extension is named after mathematician László Lovász. [9] Consider any vector such that each . Then the Lovász extension is defined as

where the expectation is over chosen from the uniform distribution on the interval . The Lovász extension is a convex function if and only if is a submodular function.

Multilinear extension

Consider any vector such that each . Then the multilinear extension is defined as [10] [11] .

Intuitively, xi represents the probability that item i is chosen for the set. For every set S, the two inner products represent the probability that the chosen set is exactly S. Therefore, the sum represents the expected value of f for the set formed by choosing each item i at random with probability xi, independently of the other items.

Convex closure

Consider any vector such that each . Then the convex closure is defined as .

The convex closure of any set function is convex over .

Concave closure

Consider any vector such that each . Then the concave closure is defined as .

Relations between continuous extensions

For the extensions discussed above, it can be shown that when is submodular. [12]

Properties

  1. The class of submodular functions is closed under non-negative linear combinations. Consider any submodular function and non-negative numbers . Then the function defined by is submodular.
  2. For any submodular function , the function defined by is submodular.
  3. The function , where is a real number, is submodular whenever is monotone submodular. More generally, is submodular, for any non decreasing concave function .
  4. Consider a random process where a set is chosen with each element in being included in independently with probability . Then the following inequality is true where is the empty set. More generally consider the following random process where a set is constructed as follows. For each of construct by including each element in independently into with probability . Furthermore let . Then the following inequality is true .[ citation needed ]

Optimization problems

Submodular functions have properties which are very similar to convex and concave functions. For this reason, an optimization problem which concerns optimizing a convex or concave function can also be described as the problem of maximizing or minimizing a submodular function subject to some constraints.

Submodular set function minimization

The hardness of minimizing a submodular set function depends on constraints imposed on the problem.

  1. The unconstrained problem of minimizing a submodular function is computable in polynomial time, [13] [14] and even in strongly-polynomial time. [15] [16] Computing the minimum cut in a graph is a special case of this minimization problem.
  2. The problem of minimizing a submodular function with a cardinality lower bound is NP-hard, with polynomial factor lower bounds on the approximation factor. [17] [18]

Submodular set function maximization

Unlike the case of minimization, maximizing a generic submodular function is NP-hard even in the unconstrained setting. Thus, most of the works in this field are concerned with polynomial-time approximation algorithms, including greedy algorithms or local search algorithms.

  1. The problem of maximizing a non-negative submodular function admits a 1/2 approximation algorithm. [19] [20] Computing the maximum cut of a graph is a special case of this problem.
  2. The problem of maximizing a monotone submodular function subject to a cardinality constraint admits a approximation algorithm. [21] [ page needed ] [22] The maximum coverage problem is a special case of this problem.
  3. The problem of maximizing a monotone submodular function subject to a matroid constraint (which subsumes the case above) also admits a approximation algorithm. [23] [24] [25]

Many of these algorithms can be unified within a semi-differential based framework of algorithms. [18]

Apart from submodular minimization and maximization, there are several other natural optimization problems related to submodular functions.

  1. Minimizing the difference between two submodular functions [26] is not only NP hard, but also inapproximable. [27]
  2. Minimization/maximization of a submodular function subject to a submodular level set constraint (also known as submodular optimization subject to submodular cover or submodular knapsack constraint) admits bounded approximation guarantees. [28]
  3. Partitioning data based on a submodular function to maximize the average welfare is known as the submodular welfare problem, which also admits bounded approximation guarantees (see welfare maximization).

Applications

Submodular functions naturally occur in several real world applications, in economics, game theory, machine learning and computer vision. [4] [29] Owing to the diminishing returns property, submodular functions naturally model costs of items, since there is often a larger discount, with an increase in the items one buys. Submodular functions model notions of complexity, similarity and cooperation when they appear in minimization problems. In maximization problems, on the other hand, they model notions of diversity, information and coverage.

See also

Citations

  1. H. Lin and J. Bilmes, A Class of Submodular Functions for Document Summarization, ACL-2011.
  2. S. Tschiatschek, R. Iyer, H. Wei and J. Bilmes, Learning Mixtures of Submodular Functions for Image Collection Summarization, NIPS-2014.
  3. A. Krause and C. Guestrin, Near-optimal nonmyopic value of information in graphical models, UAI-2005.
  4. 1 2 A. Krause and C. Guestrin, Beyond Convexity: Submodularity in Machine Learning, Tutorial at ICML-2008
  5. (Schrijver  2003 ,§44, p. 766)
  6. Buchbinder, Niv; Feldman, Moran (2018). "Submodular Functions Maximization Problems". In Gonzalez, Teofilo F. (ed.). Handbook of Approximation Algorithms and Metaheuristics, Second Edition: Methodologies and Traditional Applications. Chapman and Hall/CRC. doi:10.1201/9781351236423. ISBN   9781351236423.
  7. "Information Processing and Learning" (PDF). cmu.
  8. Fujishige (2005) p.22
  9. Lovász, L. (1983). "Submodular functions and convexity". Mathematical Programming the State of the Art. pp. 235–257. doi:10.1007/978-3-642-68874-4_10. ISBN   978-3-642-68876-8. S2CID   117358746.
  10. Vondrak, Jan (2008-05-17). "Optimal approximation for the submodular welfare problem in the value oracle model". Proceedings of the fortieth annual ACM symposium on Theory of computing. STOC '08. New York, NY, USA: Association for Computing Machinery. pp. 67–74. doi:10.1145/1374376.1374389. ISBN   978-1-60558-047-0. S2CID   170510.
  11. Calinescu, Gruia; Chekuri, Chandra; Pál, Martin; Vondrák, Jan (January 2011). "Maximizing a Monotone Submodular Function Subject to a Matroid Constraint". SIAM Journal on Computing. 40 (6): 1740–1766. doi:10.1137/080733991. ISSN   0097-5397.
  12. Vondrák, Jan. "Polyhedral techniques in combinatorial optimization: Lecture 17" (PDF).
  13. Grötschel, M.; Lovasz, L.; Schrijver, A. (1981). "The ellipsoid method and its consequences in combinatorial optimization". Combinatorica. 1 (2): 169–197. doi:10.1007/BF02579273. hdl: 10068/182482 . S2CID   43787103.
  14. Cunningham, W. H. (1985). "On submodular function minimization". Combinatorica. 5 (3): 185–192. doi:10.1007/BF02579361. S2CID   33192360.
  15. Iwata, S.; Fleischer, L.; Fujishige, S. (2001). "A combinatorial strongly polynomial algorithm for minimizing submodular functions". J. ACM. 48 (4): 761–777. doi:10.1145/502090.502096. S2CID   888513.
  16. Schrijver, A. (2000). "A combinatorial algorithm minimizing submodular functions in strongly polynomial time". J. Combin. Theory Ser. B. 80 (2): 346–355. doi: 10.1006/jctb.2000.1989 .
  17. Z. Svitkina and L. Fleischer, Submodular approximation: Sampling-based algorithms and lower bounds, SIAM Journal on Computing (2011).
  18. 1 2 R. Iyer, S. Jegelka and J. Bilmes, Fast Semidifferential based submodular function optimization, Proc. ICML (2013).
  19. U. Feige, V. Mirrokni and J. Vondrák, Maximizing non-monotone submodular functions, Proc. of 48th FOCS (2007), pp. 461–471.
  20. N. Buchbinder, M. Feldman, J. Naor and R. Schwartz, A tight linear time (1/2)-approximation for unconstrained submodular maximization, Proc. of 53rd FOCS (2012), pp. 649-658.
  21. Nemhauser, George; Wolsey, L. A.; Fisher, M. L. (1978). "An analysis of approximations for maximizing submodular set functions I". Mathematical Programming. 14 (14): 265–294. doi:10.1007/BF01588971. S2CID   206800425.
  22. Williamson, David P. "Bridging Continuous and Discrete Optimization: Lecture 23" (PDF).
  23. G. Calinescu, C. Chekuri, M. Pál and J. Vondrák, Maximizing a submodular set function subject to a matroid constraint, SIAM J. Comp. 40:6 (2011), 1740-1766.
  24. M. Feldman, J. Naor and R. Schwartz, A unified continuous greedy algorithm for submodular maximization, Proc. of 52nd FOCS (2011).
  25. Y. Filmus, J. Ward, A tight combinatorial algorithm for submodular maximization subject to a matroid constraint, Proc. of 53rd FOCS (2012), pp. 659-668.
  26. M. Narasimhan and J. Bilmes, A submodular-supermodular procedure with applications to discriminative structure learning, In Proc. UAI (2005).
  27. R. Iyer and J. Bilmes, Algorithms for Approximate Minimization of the Difference between Submodular Functions, In Proc. UAI (2012).
  28. R. Iyer and J. Bilmes, Submodular Optimization Subject to Submodular Cover and Submodular Knapsack Constraints, In Advances of NIPS (2013).
  29. J. Bilmes, Submodularity in Machine Learning Applications, Tutorial at AAAI-2015.

Related Research Articles

<span class="mw-page-title-main">Linear programming</span> Method to solve some optimization problems

Linear programming (LP), also called linear optimization, is a method to achieve the best outcome in a mathematical model whose requirements are represented by linear relationships. Linear programming is a special case of mathematical programming.

Big <i>O</i> notation Describes limiting behavior of a function

Big O notation is a mathematical notation that describes the limiting behavior of a function when the argument tends towards a particular value or infinity. Big O is a member of a family of notations invented by German mathematicians Paul Bachmann, Edmund Landau, and others, collectively called Bachmann–Landau notation or asymptotic notation. The letter O was chosen by Bachmann to stand for Ordnung, meaning the order of approximation.

<span class="mw-page-title-main">Greedy algorithm</span> Sequence of locally optimal choices

A greedy algorithm is any algorithm that follows the problem-solving heuristic of making the locally optimal choice at each stage. In many problems, a greedy strategy does not produce an optimal solution, but a greedy heuristic can yield locally optimal solutions that approximate a globally optimal solution in a reasonable amount of time.

In the mathematical field of real analysis, the monotone convergence theorem is any of a number of related theorems proving the convergence of monotonic sequences that are also bounded. Informally, the theorems state that if a sequence is increasing and bounded above by a supremum, then the sequence will converge to the supremum; in the same way, if a sequence is decreasing and is bounded below by an infimum, it will converge to the infimum.

In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou.

Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets. Many classes of convex optimization problems admit polynomial-time algorithms, whereas mathematical optimization is in general NP-hard.

In mathematics, a polymatroid is a polytope associated with a submodular function. The notion was introduced by Jack Edmonds in 1970. It is also described as the multiset analogue of the matroid.

The Frank–Wolfe algorithm is an iterative first-order optimization algorithm for constrained convex optimization. Also known as the conditional gradient method, reduced gradient algorithm and the convex combination algorithm, the method was originally proposed by Marguerite Frank and Philip Wolfe in 1956. In each iteration, the Frank–Wolfe algorithm considers a linear approximation of the objective function, and moves towards a minimizer of this linear function.

<span class="mw-page-title-main">Scoring rule</span> Measure for evaluating probabilistic forecasts

In decision theory, a scoring rule provides a summary measure for the evaluation of probabilistic predictions or forecasts. It is applicable to tasks in which predictions assign probabilities to events, i.e. one issues a probability distribution as prediction. This includes probabilistic classification of a set of mutually exclusive outcomes or classes.

Linear Programming Boosting (LPBoost) is a supervised classifier from the boosting family of classifiers. LPBoost maximizes a margin between training samples of different classes and hence also belongs to the class of margin-maximizing supervised classification algorithms. Consider a classification function

In mathematical optimization, linear-fractional programming (LFP) is a generalization of linear programming (LP). Whereas the objective function in a linear program is a linear function, the objective function in a linear-fractional program is a ratio of two linear functions. A linear program can be regarded as a special case of a linear-fractional program in which the denominator is the constant function 1.

The maximum coverage problem is a classical question in computer science, computational complexity theory, and operations research. It is a problem that is widely taught in approximation algorithms.

In mathematics, there are many kinds of inequalities involving matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with traces of matrices.

In mathematics, a subadditive set function is a set function whose value, informally, has the property that the value of function on the union of two sets is at most the sum of values of the function on each of the sets. This is thematically related to the subadditivity property of real-valued functions.

In the mathematical theory of probability, the drift-plus-penalty method is used for optimization of queueing networks and other stochastic systems.

In financial mathematics and stochastic optimization, the concept of risk measure is used to quantify the risk involved in a random outcome or risk position. Many risk measures have hitherto been proposed, each having certain characteristics. The entropic value at risk (EVaR) is a coherent risk measure introduced by Ahmadi-Javid, which is an upper bound for the value at risk (VaR) and the conditional value at risk (CVaR), obtained from the Chernoff inequality. The EVaR can also be represented by using the concept of relative entropy. Because of its connection with the VaR and the relative entropy, this risk measure is called "entropic value at risk". The EVaR was developed to tackle some computational inefficiencies of the CVaR. Getting inspiration from the dual representation of the EVaR, Ahmadi-Javid developed a wide class of coherent risk measures, called g-entropic risk measures. Both the CVaR and the EVaR are members of this class.

<span class="mw-page-title-main">Event detection for WSN</span>

Wireless sensor networks (WSN) are a spatially distributed network of autonomous sensors used for monitoring an environment. Energy cost is a major limitation for WSN requiring the need for energy efficient networks and processing. One of major energy costs in WSN is the energy spent on communication between nodes and it is sometimes desirable to only send data to a gateway node when an event of interest is triggered at a sensor. Sensors will then only open communication during a probable event, saving on communication costs. Fields interested in this type of network include surveillance, home automation, disaster relief, traffic control, health care and more.

In number theory, the prime omega functions and count the number of prime factors of a natural number Thereby counts each distinct prime factor, whereas the related function counts the total number of prime factors of honoring their multiplicity. That is, if we have a prime factorization of of the form for distinct primes , then the respective prime omega functions are given by and . These prime factor counting functions have many important number theoretic relations.

Quadratic pseudo-Boolean optimisation (QPBO) is a combinatorial optimization method for quadratic pseudo-Boolean functions in the form

The welfare maximization problem is an optimization problem studied in economics and computer science. Its goal is to partition a set of items among agents with different utility functions, such that the welfare – defined as the sum of the agents' utilities – is as high as possible. In other words, the goal is to find an item allocation satisfying the utilitarian rule.

References