Population-weighted density is an alternate metric for the population density of a region that attempts to measure the density as experienced by the average person who lives in the region.
Unlike conventional, or "area weighted" density, it is not changed when empty or extremely low-population areas are added to the region whose density is being computed.
Population-weighted density is generally computed by subdividing a region into parcels (alternately called "zones" or "subsets"), each with conventional density . Then, a mean is computed, either arithmetically, asor geometrically, asPopulation-weighted density is equivalent to area-weighted density when the parcel is the entire region, and subdividing parcels can never reduce population-weighted density. As a result, population-weighted density is always at least as high as area-weighted density for a given region.
Population-weighted density was introduced by John Craig of the UK Office of Population Censuses and Surveys in 1984. This paper introduced the formulas for computing mean population-weighted density both arithmetically and geometrically, and provided evidence that it differed from conventionally-defined density. The paper also emphasized that what was provided was not a single definition but a family of definitions, dependent on the definition of a parcel. It argues that this is in fact a positive characteristic, as it "adds to the need to think about what the fundamental unit of density actually is." In addition, it describes the problems of selecting parcels that are too large or two small, and explains the property that smaller parcels lead to larger densities. [1]
In 1998, Richardsun, Brunton, and Roddis rediscovered this technique, attempting to specifically prevent the problem of urban boundary definitions affecting density computations. Specifically, they characterize the problem as one of computing "perceived density" and provide the arithmetic mean version of the formula. They also analyze the relationship between parcel size and density value and find a linear relationship between density value and log parcel area across several different cities. [2]
In more recent times, population-weighted density has been used in analysis in both scholarly and non-scholarly sources, such as analysis of wage inequality, [3] epidemiology, [4] and economics. [5] The United States Census using population-weighted density in their 2012 report on patterns of metropolitan change, [6] leading to an increase in the metric's popularity. [7]
Population-weighted density's dependence on parcel size can often lead to criticism as lack of standardization can make population-weighted density difficult to compare across locales. Additionally, the choice of parcel to use is often poorly justified. [7]
In probability theory, the expected value is a generalization of the weighted average. Informally, the expected value is the mean of the possible values a random variable can take, weighted by the probability of those outcomes. Since it is obtained through arithmetic, the expected value sometimes may not even be included in the sample data set; it is not the value you would "expect" to get in reality.
In mathematics, the exponential function is the unique real function which maps zero to one and has a derivative equal to its value. The exponential of a variable is denoted or , with the two notations used interchangeably. It is called exponential because its argument can be seen as an exponent to which a constant number e ≈ 2.718, the base, is raised. There are several other definitions of the exponential function, which are all equivalent although being of very different nature.
In mathematics, the geometric mean is a mean or average which indicates a central tendency of a finite collection of positive real numbers by using the product of their values. The geometric mean of numbers is the nth root of their product, i.e., for a collection of numbers a1, a2, ..., an, the geometric mean is defined as
In mathematics, generalized means are a family of functions for aggregating sets of numbers. These include as special cases the Pythagorean means.
In mathematics, the harmonic mean is a kind of average, one of the Pythagorean means.
A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means in mathematics, especially in statistics. Each attempts to summarize or typify a given group of data, illustrating the magnitude and sign of the data set. Which of these measures is most illuminating depends on what is being measured, and on context and purpose.
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is
The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to 2.718281828459. The natural logarithm of x is generally written as ln x, logex, or sometimes, if the base e is implicit, simply log x. Parentheses are sometimes added for clarity, giving ln(x), loge(x), or log(x). This is done particularly when the argument to the logarithm is not a single symbol, so as to prevent ambiguity.
In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average rate; the distance parameter could be any meaningful mono-dimensional measure of the process, such as time between production errors, or length along a roll of fabric in the weaving manufacturing process. It is a particular case of the gamma distribution. It is the continuous analogue of the geometric distribution, and it has the key property of being memoryless. In addition to being used for the analysis of Poisson point processes it is found in various other contexts.
In mathematics, exponentiation, denoted bn, is an operation involving two numbers: the base, b, and the exponent or power, n. When n is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, bn is the product of multiplying n bases: In particular, .
In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).
Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:
In statistics, the weighted geometric mean is a generalization of the geometric mean using the weighted arithmetic mean.
The Theil index is a statistic primarily used to measure economic inequality and other economic phenomena, though it has also been used to measure racial segregation. The Theil index TT is the same as redundancy in information theory which is the maximum possible entropy of the data minus the observed entropy. It is a special case of the generalized entropy index. It can be viewed as a measure of redundancy, lack of diversity, isolation, segregation, inequality, non-randomness, and compressibility. It was proposed by a Dutch econometrician Henri Theil (1924–2000) at the Erasmus University Rotterdam.
A diversity index is a method of measuring how many different types there are in a dataset. Some more sophisticated indices also account for the phylogenetic relatedness among the types. Diversity indices are statistical representations of different aspects of biodiversity, which are useful simplifications for comparing different communities or sites.
In mathematics and statistics, a circular mean or angular mean is a mean designed for angles and similar cyclic quantities, such as times of day, and fractional parts of real numbers.
In probability theory and statistics, the normal-gamma distribution is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and precision.
The Widom insertion method is a statistical thermodynamic approach to the calculation of material and mixture properties. It is named for Benjamin Widom, who derived it in 1963. In general, there are two theoretical approaches to determining the statistical mechanical properties of materials. The first is the direct calculation of the overall partition function of the system, which directly yields the system free energy. The second approach, known as the Widom insertion method, instead derives from calculations centering on one molecule. The Widom insertion method directly yields the chemical potential of one component rather than the system free energy. This approach is most widely applied in molecular computer simulations but has also been applied in the development of analytical statistical mechanical models. The Widom insertion method can be understood as an application of the Jarzynski equality since it measures the excess free energy difference via the average work needed to perform, when changing the system from a state with N molecules to a state with N+1 molecules. Therefore it measures the excess chemical potential since , where .
In mathematics and statistics, the Fréchet mean is a generalization of centroids to metric spaces, giving a single representative point or central tendency for a cluster of points. It is named after Maurice Fréchet. Karcher mean is the renaming of the Riemannian Center of Mass construction developed by Karsten Grove and Hermann Karcher. On the real numbers, the arithmetic mean, median, geometric mean, and harmonic mean can all be interpreted as Fréchet means for different distance functions.
In probability theory and statistics, the Poisson binomial distribution is the discrete probability distribution of a sum of independent Bernoulli trials that are not necessarily identically distributed. The concept is named after Siméon Denis Poisson.