![]() | This article has multiple issues. Please help improve it or discuss these issues on the talk page . (Learn how and when to remove these template messages)
|
The uncertainty theory invented by Baoding Liu [1] is a branch of mathematics based on normality, monotonicity, self-duality, countable subadditivity, and product measure axioms.[ clarification needed ]
Mathematical measures of the likelihood of an event being true include probability theory, capacity, fuzzy logic, possibility, and credibility, as well as uncertainty.
Axiom 1. (Normality Axiom) .
Axiom 2. (Self-Duality Axiom) .
Axiom 3. (Countable Subadditivity Axiom) For every countable sequence of events , we have
Axiom 4. (Product Measure Axiom) Let be uncertainty spaces for . Then the product uncertain measure is an uncertain measure on the product σ-algebra satisfying
Principle. (Maximum Uncertainty Principle) For any event, if there are multiple reasonable values that an uncertain measure may take, then the value as close to 0.5 as possible is assigned to the event.
An uncertain variable is a measurable function ξ from an uncertainty space to the set of real numbers, i.e., for any Borel set B of real numbers, the set is an event.
Uncertainty distribution is inducted to describe uncertain variables.
Definition: The uncertainty distribution of an uncertain variable ξ is defined by .
Theorem (Peng and Iwamura, Sufficient and Necessary Condition for Uncertainty Distribution): A function is an uncertain distribution if and only if it is an increasing function except and .
Definition: The uncertain variables are said to be independent if
for any Borel sets of real numbers.
Theorem 1: The uncertain variables are independent if
for any Borel sets of real numbers.
Theorem 2: Let be independent uncertain variables, and measurable functions. Then are independent uncertain variables.
Theorem 3: Let be uncertainty distributions of independent uncertain variables respectively, and the joint uncertainty distribution of uncertain vector . If are independent, then we have
for any real numbers .
Theorem: Let be independent uncertain variables, and a measurable function. Then is an uncertain variable such that
where are Borel sets, and means for any.
Definition: Let be an uncertain variable. Then the expected value of is defined by
provided that at least one of the two integrals is finite.
Theorem 1: Let be an uncertain variable with uncertainty distribution . If the expected value exists, then
Theorem 2: Let be an uncertain variable with regular uncertainty distribution . If the expected value exists, then
Theorem 3: Let and be independent uncertain variables with finite expected values. Then for any real numbers and , we have
Definition: Let be an uncertain variable with finite expected value . Then the variance of is defined by
Theorem: If be an uncertain variable with finite expected value, and are real numbers, then
Definition: Let be an uncertain variable, and . Then
is called the α-optimistic value to , and
is called the α-pessimistic value to .
Theorem 1: Let be an uncertain variable with regular uncertainty distribution . Then its α-optimistic value and α-pessimistic value are
Theorem 2: Let be an uncertain variable, and . Then we have
Theorem 3: Suppose that and are independent uncertain variables, and . Then we have
,
,
,
,
,
.
Definition: Let be an uncertain variable with uncertainty distribution . Then its entropy is defined by
where .
Theorem 1(Dai and Chen): Let be an uncertain variable with regular uncertainty distribution . Then
Theorem 2: Let and be independent uncertain variables. Then for any real numbers and , we have
Theorem 3: Let be an uncertain variable whose uncertainty distribution is arbitrary but the expected value and variance . Then
Theorem 1(Liu, Markov Inequality): Let be an uncertain variable. Then for any given numbers and , we have
Theorem 2 (Liu, Chebyshev Inequality) Let be an uncertain variable whose variance exists. Then for any given number , we have
Theorem 3 (Liu, Holder's Inequality) Let and be positive numbers with , and let and be independent uncertain variables with and . Then we have
Theorem 4:(Liu [127], Minkowski Inequality) Let be a real number with , and let and be independent uncertain variables with and . Then we have
Definition 1: Suppose that are uncertain variables defined on the uncertainty space . The sequence is said to be convergent a.s. to if there exists an event with such that
for every . In that case we write ,a.s.
Definition 2: Suppose that are uncertain variables. We say that the sequence converges in measure to if
for every .
Definition 3: Suppose that are uncertain variables with finite expected values. We say that the sequence converges in mean to if
Definition 4: Suppose that are uncertainty distributions of uncertain variables , respectively. We say that the sequence converges in distribution to if at any continuity point of .
Theorem 1: Convergence in Mean Convergence in Measure Convergence in Distribution. However, Convergence in Mean Convergence Almost Surely Convergence in Distribution.
Definition 1: Let be an uncertainty space, and . Then the conditional uncertain measure of A given B is defined by
Theorem 1: Let be an uncertainty space, and B an event with . Then M{·|B} defined by Definition 1 is an uncertain measure, and is an uncertainty space.
Definition 2: Let be an uncertain variable on . A conditional uncertain variable of given B is a measurable function from the conditional uncertainty space to the set of real numbers such that
Definition 3: The conditional uncertainty distribution of an uncertain variable given B is defined by
provided that .
Theorem 2: Let be an uncertain variable with regular uncertainty distribution , and a real number with . Then the conditional uncertainty distribution of given is
Theorem 3: Let be an uncertain variable with regular uncertainty distribution , and a real number with . Then the conditional uncertainty distribution of given is
Definition 4: Let be an uncertain variable. Then the conditional expected value of given B is defined by
provided that at least one of the two integrals is finite.
In mathematical analysis, the Dirac delta function, also known as the unit impulse, is a generalized function on the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire real line is equal to one. Since there is no function having this property, to model the delta "function" rigorously involves the use of limits or, as is common in mathematics, measure theory and the theory of distributions.
In probability theory, the central limit theorem (CLT) states that, under appropriate conditions, the distribution of a normalized version of the sample mean converges to a standard normal distribution. This holds even if the original variables themselves are not normally distributed. There are several versions of the CLT, each applying in the context of different conditions.
Distributions, also known as Schwartz distributions or generalized functions, are objects that generalize the classical notion of functions in mathematical analysis. Distributions make it possible to differentiate functions whose derivatives do not exist in the classical sense. In particular, any locally integrable function has a distributional derivative.
In mathematics, a base (or basis; pl.: bases) for the topology τ of a topological space (X, τ) is a family of open subsets of X such that every open set of the topology is equal to the union of some sub-family of . For example, the set of all open intervals in the real number line is a basis for the Euclidean topology on because every open interval is an open set, and also every open subset of can be written as a union of some family of open intervals.
In mathematics, the adele ring of a global field is a central object of class field theory, a branch of algebraic number theory. It is the restricted product of all the completions of the global field and is an example of a self-dual topological ring.
In mathematics, a differential operator is an operator defined as a function of the differentiation operator. It is helpful, as a matter of notation first, to consider differentiation as an abstract operation that accepts a function and returns another function.
Vapnik–Chervonenkis theory was developed during 1960–1990 by Vladimir Vapnik and Alexey Chervonenkis. The theory is a form of computational learning theory, which attempts to explain the learning process from a statistical point of view.
In mathematics, integral equations are equations in which an unknown function appears under an integral sign. In mathematical notation, integral equations may thus be expressed as being of the form:
In the field of mathematical optimization, stochastic programming is a framework for modeling optimization problems that involve uncertainty. A stochastic program is an optimization problem in which some or all problem parameters are uncertain, but follow known probability distributions. This framework contrasts with deterministic optimization, in which all problem parameters are assumed to be known exactly. The goal of stochastic programming is to find a decision which both optimizes some criteria chosen by the decision maker, and appropriately accounts for the uncertainty of the problem parameters. Because many real-world decisions involve uncertainty, stochastic programming has found applications in a broad range of areas ranging from finance to transportation to energy optimization.
In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel, Fréchet and Weibull families also known as type I, II and III extreme value distributions. By the extreme value theorem the GEV distribution is the only possible limit distribution of properly normalized maxima of a sequence of independent and identically distributed random variables. Note that a limit distribution needs to exist, which requires regularity conditions on the tail of the distribution. Despite this, the GEV distribution is often used as an approximation to model the maxima of long (finite) sequences of random variables.
In mathematics, the theory of optimal stopping or early stopping is concerned with the problem of choosing a time to take a particular action, in order to maximise an expected reward or minimise an expected cost. Optimal stopping problems can be found in areas of statistics, economics, and mathematical finance. A key example of an optimal stopping problem is the secretary problem. Optimal stopping problems can often be written in the form of a Bellman equation, and are therefore often solved using dynamic programming.
Expected shortfall (ES) is a risk measure—a concept used in the field of financial risk measurement to evaluate the market risk or credit risk of a portfolio. The "expected shortfall at q% level" is the expected return on the portfolio in the worst of cases. ES is an alternative to value at risk that is more sensitive to the shape of the tail of the loss distribution.
In mathematics, uniform integrability is an important concept in real analysis, functional analysis and measure theory, and plays a vital role in the theory of martingales.
An -superprocess, , within mathematics probability theory is a stochastic process on that is usually constructed as a special limit of near-critical branching diffusions.
In financial mathematics, tail value at risk (TVaR), also known as tail conditional expectation (TCE) or conditional tail expectation (CTE), is a risk measure associated with the more general value at risk. It quantifies the expected value of the loss given that an event outside a given probability level has occurred.
Bilinear time–frequency distributions, or quadratic time–frequency distributions, arise in a sub-field of signal analysis and signal processing called time–frequency signal processing, and, in the statistical analysis of time series data. Such methods are used where one needs to deal with a situation where the frequency composition of a signal may be changing over time; this sub-field used to be called time–frequency signal analysis, and is now more often called time–frequency signal processing due to the progress in using these methods to a wide range of signal-processing problems.
In mathematics, the spectral theory of ordinary differential equations is the part of spectral theory concerned with the determination of the spectrum and eigenfunction expansion associated with a linear ordinary differential equation. In his dissertation, Hermann Weyl generalized the classical Sturm–Liouville theory on a finite closed interval to second order differential operators with singularities at the endpoints of the interval, possibly semi-infinite or infinite. Unlike the classical case, the spectrum may no longer consist of just a countable set of eigenvalues, but may also contain a continuous part. In this case the eigenfunction expansion involves an integral over the continuous part with respect to a spectral measure, given by the Titchmarsh–Kodaira formula. The theory was put in its final simplified form for singular differential equations of even degree by Kodaira and others, using von Neumann's spectral theorem. It has had important applications in quantum mechanics, operator theory and harmonic analysis on semisimple Lie groups.
In mathematical analysis and its applications, a function of several real variables or real multivariate function is a function with more than one argument, with all arguments being real variables. This concept extends the idea of a function of a real variable to several variables. The "input" variables take real values, while the "output", also called the "value of the function", may be real or complex. However, the study of the complex-valued functions may be easily reduced to the study of the real-valued functions, by considering the real and imaginary parts of the complex function; therefore, unless explicitly specified, only real-valued functions will be considered in this article.
This article summarizes several identities in exterior calculus.
In mathematical analysis, the spaces of test functions and distributions are topological vector spaces (TVSs) that are used in the definition and application of distributions. Test functions are usually infinitely differentiable complex-valued functions on a non-empty open subset that have compact support. The space of all test functions, denoted by is endowed with a certain topology, called the canonical LF-topology, that makes into a complete Hausdorff locally convex TVS. The strong dual space of is called the space of distributions on and is denoted by where the "" subscript indicates that the continuous dual space of denoted by is endowed with the strong dual topology.