Population weighted density

Last updated
Example of the use of population-weighted density. Both regions are 9000 units in area total and have a population of 900 total. Each is divided into 9 parcels. The area-weighted densities are then both 0.10. However, the population weighted density of the first region is greater as its population is grouped in a single parcel. Population Weighted Density Explanation.svg
Example of the use of population-weighted density. Both regions are 9000 units in area total and have a population of 900 total. Each is divided into 9 parcels. The area-weighted densities are then both 0.10. However, the population weighted density of the first region is greater as its population is grouped in a single parcel.

Population-weighted density is an alternate metric for the population density of a region that attempts to measure the density as experienced by the average person who lives in the region.

Contents

Unlike conventional, or "area weighted" density, it is not changed when empty or extremely low-population areas are added to the region whose density is being computed.

Formula and properties

Population-weighted density is generally computed by subdividing a region into parcels (alternately called "zones" or "subsets"), each with conventional density . Then, a mean is computed, either arithmetically, asor geometrically, asPopulation-weighted density is equivalent to area-weighted density when the parcel is the entire region, and subdividing parcels can never reduce population-weighted density. As a result, population-weighted density is always at least as high as area-weighted density for a given region.

History

Population-weighted density was introduced by John Craig of the UK Office of Population Censuses and Surveys in 1984. This paper introduced the formulas for computing mean population-weighted density both arithmetically and geometrically, and provided evidence that it differed from conventionally-defined density. The paper also emphasized that what was provided was not a single definition but a family of definitions, dependent on the definition of a parcel. It argues that this is in fact a positive characteristic, as it "adds to the need to think about what the fundamental unit of density actually is." In addition, it describes the problems of selecting parcels that are too large or two small, and explains the property that smaller parcels lead to larger densities. [1]

In 1998, Richardsun, Brunton, and Roddis rediscovered this technique, attempting to specifically prevent the problem of urban boundary definitions affecting density computations. Specifically, they characterize the problem as one of computing "perceived density" and provide the arithmetic mean version of the formula. They also analyze the relationship between parcel size and density value and find a linear relationship between density value and log parcel area across several different cities. [2]

In more recent times, population-weighted density has been used in analysis in both scholarly and non-scholarly sources, such as analysis of wage inequality, [3] epidemiology, [4] and economics. [5] The United States Census using population-weighted density in their 2012 report on patterns of metropolitan change, [6] leading to an increase in the metric's popularity. [7]

Criticism

Population-weighted density's dependence on parcel size can often lead to criticism as lack of standardization can make population-weighted density difficult to compare across locales. Additionally, the choice of parcel to use is often poorly justified. [7]

See also

Related Research Articles

<span class="mw-page-title-main">Expected value</span> Average value of a random variable

In probability theory, the expected value is a generalization of the weighted average. Informally, the expected value is the mean of the possible values a random variable can take, weighted by the probability of those outcomes. Since it is obtained through arithmetic, the expected value sometimes may not even be included in the sample data set; it is not the value you would "expect" to get in reality.

<span class="mw-page-title-main">Exponential function</span> Mathematical function, denoted exp(x) or e^x

In mathematics, the exponential function is the unique real function which maps zero to one and has a derivative equal to its value. The exponential of a variable is denoted or , with the two notations used interchangeably. It is called exponential because its argument can be seen as an exponent to which a constant number e ≈ 2.718, the base, is raised. There are several other definitions of the exponential function, which are all equivalent although being of very different nature.

<span class="mw-page-title-main">Geometric mean</span> N-th root of the product of n numbers

In mathematics, the geometric mean is a mean or average which indicates a central tendency of a finite collection of positive real numbers by using the product of their values. The geometric mean of numbers is the nth root of their product, i.e., for a collection of numbers a1, a2, ..., an, the geometric mean is defined as

<span class="mw-page-title-main">Generalized mean</span> N-th root of the arithmetic mean of the given numbers raised to the power n

In mathematics, generalized means are a family of functions for aggregating sets of numbers. These include as special cases the Pythagorean means.

In mathematics, the harmonic mean is a kind of average, one of the Pythagorean means.

A mean is a quantity representing the "center" of a collection of numbers and is intermediate to the extreme values of the set of numbers. There are several kinds of means in mathematics, especially in statistics. Each attempts to summarize or typify a given group of data, illustrating the magnitude and sign of the data set. Which of these measures is most illuminating depends on what is being measured, and on context and purpose.

<span class="mw-page-title-main">Normal distribution</span> Probability distribution

In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is

<span class="mw-page-title-main">Natural logarithm</span> Logarithm to the base of the mathematical constant e

The natural logarithm of a number is its logarithm to the base of the mathematical constant e, which is an irrational and transcendental number approximately equal to 2.718281828459. The natural logarithm of x is generally written as ln x, logex, or sometimes, if the base e is implicit, simply log x. Parentheses are sometimes added for clarity, giving ln(x), loge(x), or log(x). This is done particularly when the argument to the logarithm is not a single symbol, so as to prevent ambiguity.

<span class="mw-page-title-main">Exponential distribution</span> Probability distribution

In probability theory and statistics, the exponential distribution or negative exponential distribution is the probability distribution of the distance between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average rate; the distance parameter could be any meaningful mono-dimensional measure of the process, such as time between production errors, or length along a roll of fabric in the weaving manufacturing process. It is a particular case of the gamma distribution. It is the continuous analogue of the geometric distribution, and it has the key property of being memoryless. In addition to being used for the analysis of Poisson point processes it is found in various other contexts.

<span class="mw-page-title-main">Exponentiation</span> Arithmetic operation

In mathematics, exponentiation, denoted bn, is an operation involving two numbers: the base, b, and the exponent or power, n. When n is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, bn is the product of multiplying n bases: In particular, .

<span class="mw-page-title-main">Log-normal distribution</span> Probability distribution

In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).

Variational Bayesian methods are a family of techniques for approximating intractable integrals arising in Bayesian inference and machine learning. They are typically used in complex statistical models consisting of observed variables as well as unknown parameters and latent variables, with various sorts of relationships among the three types of random variables, as might be described by a graphical model. As typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:

  1. To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
  2. To derive a lower bound for the marginal likelihood of the observed data. This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data.

In statistics, the weighted geometric mean is a generalization of the geometric mean using the weighted arithmetic mean.

The Theil index is a statistic primarily used to measure economic inequality and other economic phenomena, though it has also been used to measure racial segregation. The Theil index TT is the same as redundancy in information theory which is the maximum possible entropy of the data minus the observed entropy. It is a special case of the generalized entropy index. It can be viewed as a measure of redundancy, lack of diversity, isolation, segregation, inequality, non-randomness, and compressibility. It was proposed by a Dutch econometrician Henri Theil (1924–2000) at the Erasmus University Rotterdam.

A diversity index is a method of measuring how many different types there are in a dataset. Some more sophisticated indices also account for the phylogenetic relatedness among the types. Diversity indices are statistical representations of different aspects of biodiversity, which are useful simplifications for comparing different communities or sites.

In mathematics and statistics, a circular mean or angular mean is a mean designed for angles and similar cyclic quantities, such as times of day, and fractional parts of real numbers.

In probability theory and statistics, the normal-gamma distribution is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean and precision.

The Widom insertion method is a statistical thermodynamic approach to the calculation of material and mixture properties. It is named for Benjamin Widom, who derived it in 1963. In general, there are two theoretical approaches to determining the statistical mechanical properties of materials. The first is the direct calculation of the overall partition function of the system, which directly yields the system free energy. The second approach, known as the Widom insertion method, instead derives from calculations centering on one molecule. The Widom insertion method directly yields the chemical potential of one component rather than the system free energy. This approach is most widely applied in molecular computer simulations but has also been applied in the development of analytical statistical mechanical models. The Widom insertion method can be understood as an application of the Jarzynski equality since it measures the excess free energy difference via the average work needed to perform, when changing the system from a state with N molecules to a state with N+1 molecules. Therefore it measures the excess chemical potential since , where .

In mathematics and statistics, the Fréchet mean is a generalization of centroids to metric spaces, giving a single representative point or central tendency for a cluster of points. It is named after Maurice Fréchet. Karcher mean is the renaming of the Riemannian Center of Mass construction developed by Karsten Grove and Hermann Karcher. On the real numbers, the arithmetic mean, median, geometric mean, and harmonic mean can all be interpreted as Fréchet means for different distance functions.

In probability theory and statistics, the Poisson binomial distribution is the discrete probability distribution of a sum of independent Bernoulli trials that are not necessarily identically distributed. The concept is named after Siméon Denis Poisson.

References

  1. Craig, John (1984-08-01). "Averaging population density". Demography. 21 (3): 405–412. doi:10.2307/2061168. ISSN   1533-7790. JSTOR   2061168.
  2. Richardson, A. J.; Brunton, P. J.; Roddis, S. M. (1998). "The Calculation of Perceived Residential Density". Road and Transport Research. 7 (2). ISSN   1037-5783.
  3. Wheeler, Christopher H. (2004-08-01). "Wage inequality and urban density". Journal of Economic Geography. 4 (4): 421–437. doi:10.1093/jnlecg/lbh033. ISSN   1468-2702.
  4. Baser, Onur (2021-02-01). "Population density index and its use for distribution of Covid-19: A case study using Turkish data". Health Policy. 125 (2): 148–154. doi:10.1016/j.healthpol.2020.10.003. hdl:20.500.11779/1385. ISSN   0168-8510. PMC   7550260 . PMID   33190934.
  5. "Density". Paul Krugman Blog. 2013-04-16. Retrieved 2025-01-25.
  6. Wilson, Steven G.; Plane, David A.; Mackun, Paul J.; Fischetti, Thomas R.; Goworowska, Justyna (September 2012). "Patterns of Metropolitan and Micropolitan Population Change: 2000 to 2010". Census.gov. Retrieved 2025-01-25.
  7. 1 2 Ottensmann, John R. (2021). "The Use (and Misuse) of Population-Weighted Density". SSRN Electronic Journal. doi:10.2139/ssrn.3970248. ISSN   1556-5068.