In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou.
Fatou's lemma can be used to prove the Fatou–Lebesgue theorem and Lebesgue's dominated convergence theorem.
In what follows, denotes the -algebra of Borel sets on .
Theorem — Fatou's lemma. Given a measure space and a set let be a sequence of -measurable non-negative functions . Define the function by for every . Then is -measurable, and
where the integrals and the limes inferior may be infinite.
Fatou's lemma remains true if its assumptions hold -almost everywhere. In other words, it is enough that there is a null set such that the values are non-negative for every To see this, note that the integrals appearing in Fatou's lemma are unchanged if we change each function on .
Fatou's lemma does not require the monotone convergence theorem, but the latter can be used to provide a quick and natural proof. A proof directly from the definitions of integrals is given further below.
let . Then:
Since
and infima and suprema of measurable functions are measurable we see that is measurable.
By the Monotone Convergence Theorem and property (1), the sup and integral may be interchanged:
where the last step used property (2).
To demonstrate that the monotone convergence theorem is not "hidden", the proof below does not use any properties of Lebesgue integral except those established here and the fact that the functions and are measurable.
Denote by the set of simple -measurable functions such that on .
Monotonicity —
1. Since we have
By definition of Lebesgue integral and the properties of supremum,
2. Let be the indicator function of the set It can be deduced from the definition of Lebesgue integral that
if we notice that, for every outside of Combined with the previous property, the inequality implies
3. First note that the claim holds if f is the indicator function of a set, by monotonicity of measures. By linearity, this also immediately implies the claim for simple functions.
Since any simple function supported on Sn is simple and supported on X, we must have
For the reverse, suppose g∈ SF(f) with By the above,
Now we turn to the main theorem
Step 1 — is -measurable, for every , as is .
Recall the closed intervals generate the Borel σ-algebra. Thus it suffices to show, for every , that . Now observe that
Every set on the right-hand side is from , which is closed under countable intersections. Thus the left-hand side is also a member of .
Similarly, it is enough to verify that , for every . Since the sequence pointwise non-decreases,
Step 2 — Given a simple function and a real number , define
Then , , and .
Step 2a. To prove the first claim, write s as a weighted sum of indicator functions of disjoint sets:
Then
Since the pre-image of the Borel set under the measurable function is measurable, and -algebras are closed under finite intersection and unions, the first claim follows.
Step 2b. To prove the second claim, note that, for each and every ,
Step 2c. To prove the third claim, suppose for contradiction there exists
Then , for every . Taking the limit as ,
This contradicts our initial assumption that .
Step 3 — From step 2 and monotonicity,
Step 4 — For every ,
Indeed, using the definition of , the non-negativity of , and the monotonicity of Lebesgue integral, we have
In accordance with Step 4, as the inequality becomes
Taking the limit as yields
as required.
Step 5 — To complete the proof, we apply the definition of Lebesgue integral to the inequality established in Step 4 and take into account that :
The proof is complete.
Equip the space with the Borel σ-algebra and the Lebesgue measure.
These sequences converge on pointwise (respectively uniformly) to the zero function (with zero integral), but every has integral one.
A suitable assumption concerning the negative parts of the sequence f1, f2, . . . of functions is necessary for Fatou's lemma, as the following example shows. Let S denote the half line [0,∞) with the Borel σ-algebra and the Lebesgue measure. For every natural number n define
This sequence converges uniformly on S to the zero function and the limit, 0, is reached in a finite number of steps: for every x ≥ 0, if n > x, then fn(x) = 0. However, every function fn has integral −1. Contrary to Fatou's lemma, this value is strictly less than the integral of the limit (0).
As discussed in § Extensions and variations of Fatou's lemma below, the problem is that there is no uniform integrable bound on the sequence from below, while 0 is the uniform bound from above.
Let f1, f2, . . . be a sequence of extended real-valued measurable functions defined on a measure space (S,Σ,μ). If there exists a non-negative integrable function g on S such that fn ≤ g for all n, then
Note: Here g integrable means that g is measurable and that .
We apply linearity of Lebesgue integral and Fatou's lemma to the sequence Since this sequence is defined -almost everywhere and non-negative.
Let f1, f2, . . . be a sequence of extended real-valued measurable functions defined on a measure space (S,Σ,μ). If there exists an integrable function g on S such that fn ≥ −g for all n, then
Apply Fatou's lemma to the non-negative sequence given by fn + g.
If in the previous setting the sequence f1, f2, . . . converges pointwise to a function fμ-almost everywhere on S, then
Note that f has to agree with the limit inferior of the functions fn almost everywhere, and that the values of the integrand on a set of measure zero have no influence on the value of the integral.
The last assertion also holds, if the sequence f1, f2, . . . converges in measure to a function f.
There exists a subsequence such that
Since this subsequence also converges in measure to f, there exists a further subsequence, which converges pointwise to f almost everywhere, hence the previous variation of Fatou's lemma is applicable to this subsubsequence.
In all of the above statements of Fatou's Lemma, the integration was carried out with respect to a single fixed measure μ. Suppose that μn is a sequence of measures on the measurable space (S,Σ) such that (see Convergence of measures)
Then, with fn non-negative integrable functions and f being their pointwise limit inferior, we have
Proof |
---|
We will prove something a bit stronger here. Namely, we will allow fn to converge μ-almost everywhere on a subset E of S. We seek to show that Let
Then μ(E-K)=0 and Thus, replacing E by E-K we may assume that fn converge to f pointwise on E. Next, note that for any simple function φ we have Hence, by the definition of the Lebesgue Integral, it is enough to show that if φ is any non-negative simple function less than or equal to f, then Let a be the minimum non-negative value of φ. Define We first consider the case when . We must have that μ(A) is infinite since where M is the (necessarily finite) maximum value of that φ attains. Next, we define We have that But An is a nested increasing sequence of functions and hence, by the continuity from below μ,
Thus,
At the same time, proving the claim in this case. The remaining case is when . We must have that μ(A) is finite. Denote, as above, by M the maximum value of φ and fix ε>0. Define Then An is a nested increasing sequence of sets whose union contains A. Thus, A-An is a decreasing sequence of sets with empty intersection. Since A has finite measure (this is why we needed to consider the two separate cases), Thus, there exists n such that Therefore, since there exists N such that Hence, for At the same time, Hence, Combining these inequalities gives that Hence, sending ε to 0 and taking the liminf in n, we get that completing the proof. |
In probability theory, by a change of notation, the above versions of Fatou's lemma are applicable to sequences of random variables X1, X2, . . . defined on a probability space ; the integrals turn into expectations. In addition, there is also a version for conditional expectations.
Let X1, X2, . . . be a sequence of non-negative random variables on a probability space and let be a sub-σ-algebra. Then
Note: Conditional expectation for non-negative random variables is always well defined, finite expectation is not needed.
Besides a change of notation, the proof is very similar to the one for the standard version of Fatou's lemma above, however the monotone convergence theorem for conditional expectations has to be applied.
Let X denote the limit inferior of the Xn. For every natural number k define pointwise the random variable
Then the sequence Y1, Y2, . . . is increasing and converges pointwise to X. For k ≤ n, we have Yk ≤ Xn, so that
by the monotonicity of conditional expectation, hence
because the countable union of the exceptional sets of probability zero is again a null set. Using the definition of X, its representation as pointwise limit of the Yk, the monotone convergence theorem for conditional expectations, the last inequality, and the definition of the limit inferior, it follows that almost surely
Let X1, X2, . . . be a sequence of random variables on a probability space and let be a sub-σ-algebra. If the negative parts
are uniformly integrable with respect to the conditional expectation, in the sense that, for ε > 0 there exists a c > 0 such that
then
Note: On the set where
satisfies
the left-hand side of the inequality is considered to be plus infinity. The conditional expectation of the limit inferior might not be well defined on this set, because the conditional expectation of the negative part might also be plus infinity.
Let ε > 0. Due to uniform integrability with respect to the conditional expectation, there exists a c > 0 such that
Since
where x+ := max{x,0} denotes the positive part of a real x, monotonicity of conditional expectation (or the above convention) and the standard version of Fatou's lemma for conditional expectations imply
Since
we have
hence
This implies the assertion.
In mathematics, the limit inferior and limit superior of a sequence can be thought of as limiting bounds on the sequence. They can be thought of in a similar fashion for a function. For a set, they are the infimum and supremum of the set's limit points, respectively. In general, when there are multiple objects around which a sequence, function, or set accumulates, the inferior and superior limits extract the smallest and largest of them; the type of object and the measure of size is context-dependent, but the notion of extreme limits is invariant. Limit inferior is also called infimum limit, limit infimum, liminf, inferior limit, lower limit, or inner limit; limit superior is also known as supremum limit, limit supremum, limsup, superior limit, upper limit, or outer limit.
In mathematics, the Lp spaces are function spaces defined using a natural generalization of the p-norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue, although according to the Bourbaki group they were first introduced by Frigyes Riesz.
In the mathematical field of real analysis, the monotone convergence theorem is any of a number of related theorems proving the good convergence behaviour of monotonic sequences, ie. sequences that are non-increasing, or non-decreasing. In its simplest form it says that a non decreasing bounded above sequence of real numbers converges to its smallest upper bound, its supremum. Likewise, a non increasing bounded below sequence convergences to its largest lower bound, its infimum. In particular infinite sums of non negative numbers converge to the supremum of the partial sums if and only if the partial sums are bounded.
In mathematics, the limit of a sequence of sets is a set whose elements are determined by the sequence in either of two equivalent ways: (1) by upper and lower bounds on the sequence that converge monotonically to the same set and (2) by convergence of a sequence of indicator functions which are themselves real-valued. As is the case with sequences of other objects, convergence is not necessary or even usual.
In measure theory, Lebesgue's dominated convergence theorem provides sufficient conditions under which almost everywhere convergence of a sequence of functions implies convergence in the L1 norm. Its power and utility are two of the primary theoretical advantages of Lebesgue integration over Riemann integration.
In probability theory, the conditional expectation, conditional expected value, or conditional mean of a random variable is its expected value evaluated with respect to the conditional probability distribution. If the random variable can take on only a finite number of values, the "conditions" are that the variable can only take on a subset of those values. More formally, in the case when the random variable is defined over a discrete probability space, the "conditions" are a partition of this probability space.
In mathematics, mixing is an abstract concept originating from physics: the attempt to describe the irreversible thermodynamic process of mixing in the everyday world: e.g. mixing paint, mixing drinks, industrial mixing.
In the field of mathematical analysis for the calculus of variations, Γ-convergence (Gamma-convergence) is a notion of convergence for functionals. It was introduced by Ennio De Giorgi.
In mathematics, the Fatou–Lebesgue theorem establishes a chain of inequalities relating the integrals of the limit inferior and the limit superior of a sequence of functions to the limit inferior and the limit superior of integrals of these functions. The theorem is named after Pierre Fatou and Henri Léon Lebesgue.
In mathematics, the concepts of essential infimum and essential supremum are related to the notions of infimum and supremum, but adapted to measure theory and functional analysis, where one often deals with statements that are not valid for all elements in a set, but rather almost everywhere, that is, except on a set of measure zero.
In mathematics, the Wasserstein distance or Kantorovich–Rubinstein metric is a distance function defined between probability distributions on a given metric space . It is named after Leonid Vaseršteĭn.
In mathematics, more specifically measure theory, there are various notions of the convergence of measures. For an intuitive general sense of what is meant by convergence of measures, consider a sequence of measures μn on a space, sharing a common collection of measurable sets. Such a sequence might represent an attempt to construct 'better and better' approximations to a desired measure μ that is difficult to obtain directly. The meaning of 'better and better' is subject to all the usual caveats for taking limits; for any error tolerance ε > 0 we require there be N sufficiently large for n ≥ N to ensure the 'difference' between μn and μ is smaller than ε. Various notions of convergence specify precisely what the word 'difference' should mean in that description; these notions are not equivalent to one another, and vary in strength.
In mathematics, uniform integrability is an important concept in real analysis, functional analysis and measure theory, and plays a vital role in the theory of martingales.
In real analysis and measure theory, the Vitali convergence theorem, named after the Italian mathematician Giuseppe Vitali, is a generalization of the better-known dominated convergence theorem of Henri Lebesgue. It is a characterization of the convergence in Lp in terms of convergence in measure and a condition related to uniform integrability.
In mathematical analysis, and especially in real, harmonic analysis and functional analysis, an Orlicz space is a type of function space which generalizes the Lp spaces. Like the Lp spaces, they are Banach spaces. The spaces are named for Władysław Orlicz, who was the first to define them in 1932.
In mathematics, especially measure theory, a set function is a function whose domain is a family of subsets of some given set and that (usually) takes its values in the extended real number line which consists of the real numbers and
In mathematical analysis, Lorentz spaces, introduced by George G. Lorentz in the 1950s, are generalisations of the more familiar spaces.
In mathematics and information theory, Sanov's theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution. In the language of large deviations theory, Sanov's theorem identifies the rate function for large deviations of the empirical measure of a sequence of i.i.d. random variables.
In probability theory, Kolmogorov's two-series theorem is a result about the convergence of random series. It follows from Kolmogorov's inequality and is used in one proof of the strong law of large numbers.
In mathematics, Kingman's subadditive ergodic theorem is one of several ergodic theorems. It can be seen as a generalization of Birkhoff's ergodic theorem. Intuitively, the subadditive ergodic theorem is a kind of random variable version of Fekete's lemma. As a result, it can be rephrased in the language of probability, e.g. using a sequence of random variables and expected values. The theorem is named after John Kingman.