In mathematics, the connective constant is a numerical quantity associated with self-avoiding walks on a lattice. It is studied in connection with the notion of universality in two-dimensional statistical physics models. [1] While the connective constant depends on the choice of lattice so itself is not universal (similarly to other lattice-dependent quantities such as the critical probability threshold for percolation), it is nonetheless an important quantity that appears in conjectures for universal laws. Furthermore, the mathematical techniques used to understand the connective constant, for example in the recent rigorous proof by Duminil-Copin and Smirnov that the connective constant of the hexagonal lattice has the precise value , may provide clues [2] to a possible approach for attacking other important open problems in the study of self-avoiding walks, notably the conjecture that self-avoiding walks converge in the scaling limit to the Schramm–Loewner evolution.
The connective constant is defined as follows. Let denote the number of n-step self-avoiding walks starting from a fixed origin point in the lattice. Since every n + m step self avoiding walk can be decomposed into an n-step self-avoiding walk and an m-step self-avoiding walk, it follows that . Then by applying Fekete's lemma to the logarithm of the above relation, the limit can be shown to exist. This number is called the connective constant, and clearly depends on the particular lattice chosen for the walk since does. The value of is precisely known only for two lattices, see below. For other lattices, has only been approximated numerically. It is conjectured that as n goes to infinity, where and , the critical amplitude, depend on the lattice, and the exponent , which is believed to be universal and dependent on the dimension of the lattice, is conjectured to be . [3]
Lattice | Connective constant |
---|---|
Hexagonal | |
Triangular | |
Square | |
Kagomé | |
Manhattan | |
L-lattice | |
lattice | |
lattice |
These values are taken from the 1998 Jensen–Guttmann paper [4] and a more recent paper by Jacobsen, Scullard and Guttmann. [5] The connective constant of the lattice, since each step on the hexagonal lattice corresponds to either two or three steps in it, can be expressed exactly as the largest real root of the polynomial
given the exact expression for the hexagonal lattice connective constant. More information about these lattices can be found in the percolation threshold article.
In 2010, Hugo Duminil-Copin and Stanislav Smirnov published the first rigorous proof of the fact that for the hexagonal lattice. [2] This had been conjectured by Nienhuis in 1982 as part of a larger study of O(n) models using renormalization techniques. [6] The rigorous proof of this fact came from a program of applying tools from complex analysis to discrete probabilistic models that has also produced impressive results about the Ising model among others. [7] The argument relies on the existence of a parafermionic observable that satisfies half of the discrete Cauchy–Riemann equations for the hexagonal lattice. We modify slightly the definition of a self-avoiding walk by having it start and end on mid-edges between vertices. Let H be the set of all mid-edges of the hexagonal lattice. For a self-avoiding walk between two mid-edges and , we define to be the number of vertices visited and its winding as the total rotation of the direction in radians when is traversed from to . The aim of the proof is to show that the partition function
converges for and diverges for where the critical parameter is given by . This immediately implies that .
Given a domain in the hexagonal lattice, a starting mid-edge , and two parameters and , we define the parafermionic observable
If and , then for any vertex in , we have
where are the mid-edges emanating from . This lemma establishes that the parafermionic observable is divergence-free. It has not been shown to be curl-free, but this would solve several open problems (see conjectures). The proof of this lemma is a clever computation that relies heavily on the geometry of the hexagonal lattice.
Next, we focus on a finite trapezoidal domain with 2L cells forming the left hand side, T cells across, and upper and lower sides at an angle of . (Picture needed.) We embed the hexagonal lattice in the complex plane so that the edge lengths are 1 and the mid-edge in the center of the left hand side is positioned at −1/2. Then the vertices in are given by
We now define partition functions for self-avoiding walks starting at and ending on different parts of the boundary. Let denote the left hand boundary, the right hand boundary, the upper boundary, and the lower boundary. Let
By summing the identity
over all vertices in and noting that the winding is fixed depending on which part of the boundary the path terminates at, we can arrive at the relation
after another clever computation. Letting , we get a strip domain and partition functions
It was later shown that , but we do not need this for the proof. [8] We are left with the relation
From here, we can derive the inequality
And arrive by induction at a strictly positive lower bound for . Since , we have established that .
For the reverse inequality, for an arbitrary self avoiding walk on the honeycomb lattice, we perform a canonical decomposition due to Hammersley and Welsh of the walk into bridges of widths and . Note that we can bound
which implies . Finally, it is possible to bound the partition function by the bridge partition functions
And so, we have that as desired.
Nienhuis argued in favor of Flory's prediction that the mean squared displacement of the self-avoiding random walk satisfies the scaling relation , with . [2] The scaling exponent and the universal constant could be computed if the self-avoiding walk possesses a conformally invariant scaling limit, conjectured to be a Schramm–Loewner evolution with . [9]
In particle physics, the electroweak interaction or electroweak force is the unified description of two of the four known fundamental interactions of nature: electromagnetism (electromagnetic interaction) and the weak interaction. Although these two forces appear very different at everyday low energies, the theory models them as two different aspects of the same force. Above the unification energy, on the order of 246 GeV, they would merge into a single force. Thus, if the temperature is high enough – approximately 1015 K – then the electromagnetic force and weak force merge into a combined electroweak force. During the quark epoch (shortly after the Big Bang), the electroweak force split into the electromagnetic and weak force. It is thought that the required temperature of 1015 K has not been seen widely throughout the universe since before the quark epoch, and currently the highest human-made temperature in thermal equilibrium is around 5.5×1012 K (from the Large Hadron Collider).
In mathematical physics and mathematics, the Pauli matrices are a set of three 2 × 2 complex matrices that are traceless, Hermitian, involutory and unitary. Usually indicated by the Greek letter sigma, they are occasionally denoted by tau when used in connection with isospin symmetries.
In physics, a wave vector is a vector used in describing a wave, with a typical unit being cycle per metre. It has a magnitude and direction. Its magnitude is the wavenumber of the wave, and its direction is perpendicular to the wavefront. In isotropic media, this is also the direction of wave propagation.
In mathematics, the associated Legendre polynomials are the canonical solutions of the general Legendre equation
In probability theory, a distribution is said to be stable if a linear combination of two independent random variables with this distribution has the same distribution, up to location and scale parameters. A random variable is said to be stable if its distribution is stable. The stable distribution family is also sometimes referred to as the Lévy alpha-stable distribution, after Paul Lévy, the first mathematician to have studied it.
Directional statistics is the subdiscipline of statistics that deals with directions, axes or rotations in Rn. More generally, directional statistics deals with observations on compact Riemannian manifolds including the Stiefel manifold.
In probability and statistics, a circular distribution or polar distribution is a probability distribution of a random variable whose values are angles, usually taken to be in the range [0, 2π). A circular distribution is often a continuous probability distribution, and hence has a probability density, but such distributions can also be discrete, in which case they are called circular lattice distributions. Circular distributions can be used even when the variables concerned are not explicitly angles: the main consideration is that there is not usually any real distinction between events occurring at the opposite ends of the range, and the division of the range could notionally be made at any point.
In probability theory and statistics, the chi distribution is a continuous probability distribution over the non-negative real line. It is the distribution of the positive square root of a sum of squared independent Gaussian random variables. Equivalently, it is the distribution of the Euclidean distance between a multivariate Gaussian random variable and the origin. The chi distribution describes the positive square roots of a variable obeying a chi-squared distribution.
In mathematics, the von Mangoldt function is an arithmetic function named after German mathematician Hans von Mangoldt. It is an example of an important arithmetic function that is neither multiplicative nor additive.
The noncentral t-distribution generalizes Student's t-distribution using a noncentrality parameter. Whereas the central probability distribution describes how a test statistic t is distributed when the difference tested is null, the noncentral distribution describes how t is distributed when the null is false. This leads to its use in statistics, especially calculating statistical power. The noncentral t-distribution is also known as the singly noncentral t-distribution, and in addition to its primary use in statistical inference, is also used in robust modeling for data.
A theoretical motivation for general relativity, including the motivation for the geodesic equation and the Einstein field equation, can be obtained from special relativity by examining the dynamics of particles in circular orbits about the Earth. A key advantage in examining circular orbits is that it is possible to know the solution of the Einstein Field Equation a priori. This provides a means to inform and verify the formalism.
In mathematics, the Schur orthogonality relations, which were proven by Issai Schur through Schur's lemma, express a central fact about representations of finite groups. They admit a generalization to the case of compact groups in general, and in particular compact Lie groups, such as the rotation group SO(3).
A hydrogen-like atom (or hydrogenic atom) is any atom or ion with a single valence electron. These atoms are isoelectronic with hydrogen. Examples of hydrogen-like atoms include, but are not limited to, hydrogen itself, all alkali metals such as Rb and Cs, singly ionized alkaline earth metals such as Ca+ and Sr+ and other ions such as He+, Li2+, and Be3+ and isotopes of any of the above. A hydrogen-like atom includes a positively charged core consisting of the atomic nucleus and any core electrons as well as a single valence electron. Because helium is common in the universe, the spectroscopy of singly ionized helium is important in EUV astronomy, for example, of DO white dwarf stars.
In physics and mathematics, the solid harmonics are solutions of the Laplace equation in spherical polar coordinates, assumed to be (smooth) functions . There are two kinds: the regular solid harmonics, which are well-defined at the origin and the irregular solid harmonics, which are singular at the origin. Both sets of functions play an important role in potential theory, and are obtained by rescaling spherical harmonics appropriately:
In mathematics, the Fortuin–Kasteleyn–Ginibre (FKG) inequality is a correlation inequality, a fundamental tool in statistical mechanics and probabilistic combinatorics, due to Cees M. Fortuin, Pieter W. Kasteleyn, and Jean Ginibre. Informally, it says that in many random systems, increasing events are positively correlated, while an increasing and a decreasing event are negatively correlated. It was obtained by studying the random cluster model.
In probability theory and directional statistics, a wrapped Lévy distribution is a wrapped probability distribution that results from the "wrapping" of the Lévy distribution around the unit circle.
In mathematical physics, the Belinfante–Rosenfeld tensor is a modification of the stress–energy tensor that is constructed from the canonical stress–energy tensor and the spin current so as to be symmetric yet still conserved.
Batch normalization is a method used to make training of artificial neural networks faster and more stable through normalization of the layers' inputs by re-centering and re-scaling. It was proposed by Sergey Ioffe and Christian Szegedy in 2015.
The convolutional sparse coding paradigm is an extension of the global sparse coding model, in which a redundant dictionary is modeled as a concatenation of circulant matrices. While the global sparsity constraint describes signal as a linear combination of a few atoms in the redundant dictionary , usually expressed as for a sparse vector , the alternative dictionary structure adopted by the convolutional sparse coding model allows the sparsity prior to be applied locally instead of globally: independent patches of are generated by "local" dictionaries operating over stripes of .