Tensor product model transformation

Last updated

In mathematics, the tensor product (TP) model transformation was proposed by Baranyi and Yam [1] [2] [3] [4] [5] as key concept for higher-order singular value decomposition of functions. It transforms a function (which can be given via closed formulas or neural networks, fuzzy logic, etc.) into TP function form if such a transformation is possible. If an exact transformation is not possible, then the method determines a TP function that is an approximation of the given function. Hence, the TP model transformation can provide a trade-off between approximation accuracy and complexity. [6]

Contents

A free MATLAB implementation of the TP model transformation can be downloaded at or an old version of the toolbox is available at MATLAB Central . A key underpinning of the transformation is the higher-order singular value decomposition. [7]

Besides being a transformation of functions, the TP model transformation is also a new concept in qLPV based control which plays a central role in the providing a valuable means of bridging between identification and polytopic systems theories. The TP model transformation is uniquely effective in manipulating the convex hull of polytopic forms, and, as a result has revealed and proved the fact that convex hull manipulation is a necessary and crucial step in achieving optimal solutions and decreasing conservativeness [8] [9] [2] in modern LMI based control theory. Thus, although it is a transformation in a mathematical sense, it has established a conceptually new direction in control theory and has laid the ground for further new approaches towards optimality. Further details on the control theoretical aspects of the TP model transformation can be found here: TP model transformation in control theory.

The TP model transformation motivated the definition of the "HOSVD canonical form of TP functions", [10] on which further information can be found here. It has been proved that the TP model transformation is capable of numerically reconstructing this HOSVD based canonical form. [11] Thus, the TP model transformation can be viewed as a numerical method to compute the HOSVD of functions, which provides exact results if the given function has a TP function structure and approximative results otherwise.

The TP model transformation has recently been extended in order to derive various types of convex TP functions and to manipulate them. [3] This feature has led to new optimization approaches in qLPV system analysis and design, as described at TP model transformation in control theory.

Definitions

Finite element TP function
A given function , where , is a TP function if it has the structure:

that is, using compact tensor notation (using the tensor product operation of [7] ):

where core tensor is constructed from , and row vector contains continuous univariate weighting functions . The function is the -th weighting function defined on the -th dimension, and is the -the element of vector . Finite element means that is bounded for all . For qLPV modelling and control applications a higher structure of TP functions are referred to as TP model.

Finite element TP model (TP model in short)
This is a higher structure of TP function:

Here is a tensor as , thus the size of the core tensor is . The product operator has the same role as , but expresses the fact that the tensor product is applied on the sized tensor elements of the core tensor . Vector is an element of the closed hypercube .

Finite element convex TP function or model
A TP function or model is convex if the weighting functions hold:
and

This means that is inside the convex hull defined by the core tensor for all .

TP model transformation
Assume a given TP model , where , whose TP structure maybe unknown (e.g. it is given by neural networks). The TP model transformation determines its TP structure as
,

namely it generates the core tensor and the weighting functions for all . Its free MATLAB implementation is downloadable at or at MATLAB Central .

If the given does not have TP structure (i.e. it is not in the class of TP models), then the TP model transformation determines its approximation: [6]

where trade-off is offered by the TP model transformation between complexity (number of components in the core tensor or the number of weighting functions) and the approximation accuracy. The TP model can be generated according to various constrains. Typical TP models generated by the TP model transformation are:

Properties of the TP model transformation

  • the number of weighting functions are minimized per dimensions (hence the size of the core tensor);
  • the weighting functions are one variable functions of the parameter vector in an orthonormed system for each parameter (singular functions);
  • the sub tensors of the core tensor are also in orthogonal positions;
  • the core tensor and the weighting functions are ordered according to the higher-order singular values of the parameter vector;
  • it has a unique form (except for some special cases such as there are equal singular values);
  • introduces and defines the rank of the TP function by the dimensions of the parameter vector;

Related Research Articles

Continuum mechanics is a branch of mechanics that deals with the deformation of and transmission of forces through materials modeled as a continuous mass rather than as discrete particles. The French mathematician Augustin-Louis Cauchy was the first to formulate such models in the 19th century.

In mathematics, a product is the result of multiplication, or an expression that identifies objects to be multiplied, called factors. For example, 21 is the product of 3 and 7, and is the product of and . When one factor is an integer, the product is called a multiple.

<span class="mw-page-title-main">Tensor</span> Algebraic object with geometric applications

In mathematics, a tensor is an algebraic object that describes a multilinear relationship between sets of algebraic objects related to a vector space. Tensors may map between different objects such as vectors, scalars, and even other tensors. There are many types of tensors, including scalars and vectors, dual vectors, multilinear maps between vector spaces, and even some operations such as the dot product. Tensors are defined independent of any basis, although they are often referred to by their components in a basis related to a particular coordinate system; those components form an array, which can be thought of as a high-dimensional matrix.

<span class="mw-page-title-main">Affine transformation</span> Geometric transformation that preserves lines but not angles nor the origin

In Euclidean geometry, an affine transformation or affinity is a geometric transformation that preserves lines and parallelism, but not necessarily Euclidean distances and angles.

In mathematics, the Lp spaces are function spaces defined using a natural generalization of the p-norm for finite-dimensional vector spaces. They are sometimes called Lebesgue spaces, named after Henri Lebesgue, although according to the Bourbaki group they were first introduced by Frigyes Riesz.

In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statistical model, the observed data is most probable. The point in the parameter space that maximizes the likelihood function is called the maximum likelihood estimate. The logic of maximum likelihood is both intuitive and flexible, and as such the method has become a dominant means of statistical inference.

<span class="mw-page-title-main">Vapnik–Chervonenkis theory</span> Branch of statistical computational learning theory

Vapnik–Chervonenkis theory was developed during 1960–1990 by Vladimir Vapnik and Alexey Chervonenkis. The theory is a form of computational learning theory, which attempts to explain the learning process from a statistical point of view.

In functional analysis, it is often convenient to define a linear transformation on a complete, normed vector space by first defining a linear transformation on a dense subset of and then continuously extending to the whole space via the theorem below. The resulting extension remains linear and bounded, and is thus continuous, which makes it a continuous linear extension.

Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets. Many classes of convex optimization problems admit polynomial-time algorithms, whereas mathematical optimization is in general NP-hard.

In mathematical optimization, the Karush–Kuhn–Tucker (KKT) conditions, also known as the Kuhn–Tucker conditions, are first derivative tests for a solution in nonlinear programming to be optimal, provided that some regularity conditions are satisfied.

<span class="mw-page-title-main">Kernel method</span> Class of algorithms for pattern analysis

In machine learning, kernel machines are a class of algorithms for pattern analysis, whose best known member is the support-vector machine (SVM). These methods involve using linear classifiers to solve nonlinear problems. The general task of pattern analysis is to find and study general types of relations in datasets. For many algorithms that solve these tasks, the data in raw representation have to be explicitly transformed into feature vector representations via a user-specified feature map: in contrast, kernel methods require only a user-specified kernel, i.e., a similarity function over all pairs of data points computed using inner products. The feature map in kernel machines is infinite dimensional but only requires a finite dimensional matrix from user-input according to the Representer theorem. Kernel machines are slow to compute for datasets larger than a couple of thousand examples without parallel processing.

<span class="mw-page-title-main">Scoring rule</span> Measure for evaluating probabilistic forecasts

In decision theory, a scoring rule provides a summary measure for the evaluation of probabilistic predictions or forecasts. It is applicable to tasks in which predictions assign probabilities to events, i.e. one issues a probability distribution as prediction. This includes probabilistic classification of a set of mutually exclusive outcomes or classes.

In applied mathematics, polyharmonic splines are used for function approximation and data interpolation. They are very useful for interpolating and fitting scattered data in many dimensions. Special cases include thin plate splines and natural cubic splines in one dimension.

In multilinear algebra, the tensor rank decomposition or the decomposition of a tensor is the decomposition of a tensor in terms of a sum of minimum tensors. This is an open problem.

In multilinear algebra, the higher-order singular value decomposition (HOSVD) of a tensor is a specific orthogonal Tucker decomposition. It may be regarded as one type of generalization of the matrix singular value decomposition. It has applications in computer vision, computer graphics, machine learning, scientific computing, and signal processing. Some aspects can be traced as far back as F. L. Hitchcock in 1928, but it was L. R. Tucker who developed for third-order tensors the general Tucker decomposition in the 1960s, further advocated by L. De Lathauwer et al. in their Multilinear SVD work that employs the power method, or advocated by Vasilescu and Terzopoulos that developed M-mode SVD a parallel algorithm that employs the matrix SVD.

<span class="mw-page-title-main">Logit-normal distribution</span>

In probability theory, a logit-normal distribution is a probability distribution of a random variable whose logit has a normal distribution. If Y is a random variable with a normal distribution, and t is the standard logistic function, then X = t(Y) has a logit-normal distribution; likewise, if X is logit-normally distributed, then Y = logit(X)= log (X/(1-X)) is normally distributed. It is also known as the logistic normal distribution, which often refers to a multinomial logit version (e.g.).

Within bayesian statistics for machine learning, kernel methods arise from the assumption of an inner product space or similarity structure on inputs. For some such methods, such as support vector machines (SVMs), the original formulation and its regularization were not Bayesian in nature. It is helpful to understand them from a Bayesian perspective. Because the kernels are not necessarily positive semidefinite, the underlying structure may not be inner product spaces, but instead more general reproducing kernel Hilbert spaces. In Bayesian probability kernel methods are a key component of Gaussian processes, where the kernel function is known as the covariance function. Kernel methods have traditionally been used in supervised learning problems where the input space is usually a space of vectors while the output space is a space of scalars. More recently these methods have been extended to problems that deal with multiple outputs such as in multi-task learning.

Baranyi and Yam proposed the TP model transformation as a new concept in quasi-LPV (qLPV) based control, which plays a central role in the highly desirable bridging between identification and polytopic systems theories. It is also used as a TS (Takagi-Sugeno) fuzzy model transformation. It is uniquely effective in manipulating the convex hull of polytopic forms, and, hence, has revealed and proved the fact that convex hull manipulation is a necessary and crucial step in achieving optimal solutions and decreasing conservativeness in modern linear matrix inequality based control theory. Thus, although it is a transformation in a mathematical sense, it has established a conceptually new direction in control theory and has laid the ground for further new approaches towards optimality.

Based on the key idea of higher-order singular value decomposition (HOSVD) in tensor algebra, Baranyi and Yam proposed the concept of HOSVD-based canonical form of TP functions and quasi-LPV system models. Szeidl et al. proved that the TP model transformation is capable of numerically reconstructing this canonical form.

Lagrangian field theory is a formalism in classical field theory. It is the field-theoretic analogue of Lagrangian mechanics. Lagrangian mechanics is used to analyze the motion of a system of discrete particles each with a finite number of degrees of freedom. Lagrangian field theory applies to continua and fields, which have an infinite number of degrees of freedom.

References

  1. 1 2 P. Baranyi (April 2004). "TP model transformation as a way to LMI based controller design". IEEE Transactions on Industrial Electronics. 51 (2): 387–400. doi:10.1109/tie.2003.822037. S2CID   7957799.
  2. 1 2 Baranyi, Péter (2016). TP-Model Transformation-Based-Control Design Frameworks. doi:10.1007/978-3-319-19605-3. ISBN   978-3-319-19604-6.
  3. 1 2 Baranyi, Peter (2014). "The Generalized TP Model Transformation for T–S Fuzzy Model Manipulation and Generalized Stability Verification". IEEE Transactions on Fuzzy Systems. 22 (4): 934–948. doi: 10.1109/TFUZZ.2013.2278982 .
  4. 1 2 P. Baranyi; D. Tikk; Y. Yam; R. J. Patton (2003). "From Differential Equations to PDC Controller Design via Numerical Transformation". Computers in Industry. 51 (3): 281–297. doi:10.1016/s0166-3615(03)00058-7.
  5. P. Baranyi; Y. Yam & P. Várlaki (2013). Tensor Product model transformation in polytopic model-based control. Boca Raton FL: Taylor & Francis. p. 240. ISBN   978-1-43-981816-9.
  6. 1 2 3 D. Tikk; P. Baranyi; R. J. Patton (2007). "Approximation Properties of TP Model Forms and its Consequences to TPDC Design Framework". Asian Journal of Control. 9 (3): 221–331. doi:10.1111/j.1934-6093.2007.tb00410.x. S2CID   121716136.
  7. 1 2 Lieven De Lathauwer; Bart De Moor; Joos Vandewalle (2000). "A Multilinear Singular Value Decomposition". Journal on Matrix Analysis and Applications. 21 (4): 1253–1278. CiteSeerX   10.1.1.3.4043 . doi:10.1137/s0895479896305696.
  8. A.Szollosi, and Baranyi, P. (2016). Influence of the Tensor Product model representation of qLPV models on the feasibility of Linear Matrix Inequality. Asian Journal of Control, 18(4), 1328-1342
  9. A. Szöllősi and P. Baranyi: „Improved control performance of the 3‐DoF aeroelastic wing section: a TP model based 2D parametric control performance optimization.” in Asian Journal of Control, 19(2), 450-466. / 2017
  10. 1 2 P. Baranyi; L. Szeidl; P. Várlaki; Y. Yam (July 3–5, 2006). Definition of the HOSVD-based canonical form of polytopic dynamic models. 3rd International Conference on Mechatronics (ICM 2006). Budapest, Hungary. pp. 660–665.
  11. 1 2 L. Szeidl & P. Várlaki (2009). "HOSVD Based Canonical Form for Polytopic Models of Dynamic Systems". Journal of Advanced Computational Intelligence and Intelligent Informatics. 13 (1): 52–60. doi: 10.20965/jaciii.2009.p0052 .

Baranyi, P. (2018). Extension of the Multi-TP Model Transformation to Functions with Different Numbers of Variables. Complexity, 2018.