Configural frequency analysis

Last updated

Configural frequency analysis (CFA) is a method of exploratory data analysis, introduced by Gustav A. Lienert in 1969. [1] The goal of a configural frequency analysis is to detect patterns in the data that occur significantly more (such patterns are called Types) or significantly less often (such patterns are called Antitypes) than expected by chance. Thus, the idea of a CFA is to provide by the identified types and antitypes some insight into the structure of the data. Types are interpreted as concepts which are constituted by a pattern of variable values. Antitypes are interpreted as patterns of variable values that do in general not occur together.

Contents

Basic idea of the CFA algorithm

We explain the basic idea of CFA by a simple example. Assume that we have a data set that describes for each of n patients if they show certain symptoms s1, ..., sm. We assume for simplicity that a symptom is shown or not, i.e. we have a dichotomous data set.

Each record in the data set is thus an m-tuple (x1, ..., xm) where each xi is either equal to 0 (patient does not show symptom i) or 1 (patient does show symptom i). Each such m-tuple is called a configuration. Let C be the set of all possible configurations, i.e. the set of all possible m-tuples on {0,1}m. The data set can thus be described by listing the observed frequencies f(c) of all possible configurations in C.

The basic idea of CFA is to estimate the frequency of each configuration under the assumption that the m symptoms are statistically independent. Let e(c) be this estimated frequency under the assumption of independence.

Let pi(1) be the probability that a member of the investigated population shows symptom si and pi(0) be the probability that a member of the investigated population does not show symptom si. Under the assumption that all symptoms are independent we can calculate the expected relative frequency of a configuration c = (c1 , ..., cm) by:

Now f(c) and e(c) can be compared by a statistical test (typical tests applied in CFA are Pearson's chi-squared test, the binomial test or the hypergeometric test of Lehmacher).

If the statistical test suggests for a given -level that the difference between f(c) and e(c) is significant then c is called a type if f(c) > e(c) and is called an antitype if f(c) < e(c). If there is no significant difference between f(c) and e(c), then c is neither a type nor an antitype. Thus, each configuration c can have in principle three different states. It can be a type, an antitype, or not classified.

Types and antitypes are defined symmetrically. But in practical applications researchers are mainly interested to detect types. For example, clinical studies are typically interested to detect symptom combinations that are indicators for a disease. These are by definition symptom combinations which occur more often than expected by chance, i.e. types.

Control of the alpha level

Since in CFA a significance test is applied in parallel for each configuration c there is a high risk to commit a type I error (i.e. to detect a type or antitype when the null hypothesis is true). The currently most popular method to control this is to use the Bonferroni correction for the α-level. [2] There are a number of alternative methods to control the α-level. One alternative, the Holm–Bonferroni method introduced by Sture Holm, considers the number of tests already finished when the ith test is performed. [3] Thus, in this method the alpha–level is not constant for all tests.

Algorithm in the non-dichotomous case

In our example above we assumed for simplicity that the symptoms are dichotomous. This is however not a necessary restriction. CFA can also be applied for symptoms (or more general attributes of an object) that are not dichotomous but have a finite number of degrees. In this case a configuration is an element of C = S1 x ... x Sm, where Si is the set of the possible degrees for symptom si. [2] [4] [5] [6]

Chance model

The assumption of the independence of the symptoms can be replaced by another method to calculate the expected frequencies e(c) of the configurations. Such a method is called a chance model.

In most applications of CFA the assumption that all symptoms are independent is used as the chance model. A CFA using that chance model is called first-order CFA. This is the classical method of CFA that is in many publications even considered to be the only CFA method. An example of an alternative chance model is the assumption that all configurations have the same probability. A CFA using that chance model is called zero-order CFA.

Related Research Articles

A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters.

Nonparametric statistics is the branch of statistics that is not based solely on parametrized families of probability distributions. Nonparametric statistics is based on either being distribution-free or having a specified distribution but with the distribution's parameters unspecified. Nonparametric statistics includes both descriptive statistics and statistical inference. Nonparametric tests are often used when the assumptions of parametric tests are violated.

In statistics, hypotheses suggested by a given dataset, when tested with the same dataset that suggested them, are likely to be accepted even when they are not true. This is because circular reasoning would be involved: something seems true in the limited data set; therefore we hypothesize that it is true in general; therefore we wrongly test it on the same, limited data set, which seems to confirm that it is true. Generating hypotheses based on data already observed, in the absence of testing them on new data, is referred to as post hoc theorizing.

In psychometrics, item response theory (IRT) is a paradigm for the design, analysis, and scoring of tests, questionnaires, and similar instruments measuring abilities, attitudes, or other variables. It is a theory of testing based on the relationship between individuals' performances on a test item and the test takers' levels of performance on an overall measure of the ability that item was designed to measure. Several different statistical models are used to represent both item and test taker characteristics. Unlike simpler alternatives for creating scales and evaluating questionnaire responses, it does not assume that each item is equally difficult. This distinguishes IRT from, for instance, Likert scaling, in which "All items are assumed to be replications of each other or in other words items are considered to be parallel instruments" (p. 197). By contrast, item response theory treats the difficulty of each item as information to be incorporated in scaling items.

Haplotype Group of genes from one parent

A haplotype is a group of alleles in an organism that are inherited together from a single parent.

The general linear model or general multivariate regression model is a compact way of simultaneously writing several multiple linear regression models. In that sense it is not a separate statistical linear model. The various multiple linear regression models may be compactly written as

Linear discriminant analysis (LDA), normal discriminant analysis (NDA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.

The Friedman test is a non-parametric statistical test developed by Milton Friedman. Similar to the parametric repeated measures ANOVA, it is used to detect differences in treatments across multiple test attempts. The procedure involves ranking each row together, then considering the values of ranks by columns. Applicable to complete block designs, it is thus a special case of the Durbin test.

In statistics, a latent class model (LCM) relates a set of observed multivariate variables to a set of latent variables. It is a type of latent variable model. It is called a latent class model because the latent variable is discrete. A class is characterized by a pattern of conditional probabilities that indicate the chance that variables take on certain values.

In statistics, family-wise error rate (FWER) is the probability of making one or more false discoveries, or type I errors when performing multiple hypotheses tests.

Chi-square automatic interaction detection (CHAID) is a decision tree technique based on adjusted significance testing. The technique was developed in South Africa and was published in 1980 by Gordon V. Kass, who had completed a PhD thesis on this topic. CHAID can be used for prediction as well as classification, and for detection of interaction between variables. CHAID is based on a formal extension of AID and THAID procedures of the 1960s and 1970s, which in turn were extensions of earlier research, including that performed by Belson in the UK in the 1950s. A history of earlier supervised tree methods, a detailed description of the original CHAID algorithm and the exhaustive CHAID extension by Biggs, De Ville, and Suen, is described by Gilbert.

In statistics, the Bonferroni correction is a method to counteract the multiple comparisons problem. Bonferroni correction is the simplest method for counteracting this; however, it is a conservative method that gives greater risk of failure to reject a false null hypothesis than other methods, as it ignores potentially valuable information, such as the distribution of p-values across all comparisons.

Multiple comparisons problem Problem where one considers a set of inferences simultaneously based on the observed values

In statistics, the multiple comparisons, multiplicity or multiple testing problem occurs when one considers a set of statistical inferences simultaneously or infers a subset of parameters selected based on the observed values.

In statistics, the Holm–Bonferroni method, also called the Holm method or Bonferroni–Holm method, is used to counteract the problem of multiple comparisons. It is intended to control the family-wise error rate and offers a simple test uniformly more powerful than the Bonferroni correction. It is named after Sture Holm, who codified the method, and Carlo Emilio Bonferroni.

Item tree analysis (ITA) is a data analytical method which allows constructing a hierarchical structure on the items of a questionnaire or test from observed response patterns.
Assume that we have a questionnaire with m items and that subjects can answer positive (1) or negative (0) to each of these items, i.e. the items are dichotomous. If n subjects answer the items this results in a binary data matrix D with m columns and n rows. Typical examples of this data format are test items which can be solved (1) or failed (0) by subjects. Other typical examples are questionnaires where the items are statements to which subjects can agree (1) or disagree (0).
Depending on the content of the items it is possible that the response of a subject to an item j determines her or his responses to other items. It is, for example, possible that each subject who agrees to item j will also agree to item i. In this case we say that item j implies item i. The goal of an ITA is to uncover such deterministic implications from the data set D.

In statistics, qualitative comparative analysis (QCA) is a data analysis based on set theory to examine the relationship of conditions to outcome. QCA describes the relationship in terms of necessary conditions and sufficient conditions. The technique was originally developed by Charles Ragin in 1987 to study data sets that are too small for linear regression analysis but large for cross-case analysis.

Mark Stemmler

Mark Stemmler was born on August 7, 1960 in Norwood, Massachusetts, United States. He was Professor of Psychological Methodology and Quality Assurance at the Faculty of Psychology and Sports Science, Bielefeld University from 2007 to 2011. He was also a member of the Center for Statistics at Bielefeld University. Currently he is Professor of Psychological Assessment at the Department of Psychology and Sports Science at the University of Erlangen-Nuremberg. Since 2010 he is also Adjunct Professor at the College of Health and Human Development at the Pennsylvania State University.

Anton Formann

Anton K. Formann was an Austrian research psychologist, statistician, and psychometrician. He is renowned for his contributions to item response theory, latent class analysis, the measurement of change, mixture models, categorical data analysis, and quantitative methods for research synthesis (meta-analysis).

In statistics, a false coverage rate (FCR) is the average rate of false coverage, i.e. not covering the true parameters, among the selected intervals.

Alexander von Eye is a German-American psychologist and former Professor of Methods in Psychology at the University of Vienna in Vienna, Austria. Before joining the University of Vienna in 2012, he taught at Michigan State University, where he served as chair of the Unit of Developmental Psychology from 2003 to 2008. Before joining Michigan State University in 1993, he served as Professor of Human Development and Psychology at Penn State University. He has developed methods for analyzing categorical and longitudinal data in psychology. He is a fellow of the American Psychological Association and the American Psychological Society. As of 2015, he lived in Montpellier, France.

References

  1. Lienert, G. A. (1969). "Die Konfigurationsfrequenzanalyse als Klassifikationsmethode in der klinischen Psychologie" [Configural frequency analysis as a classification method in clinical psychology]. In Irle, M. (ed.). Bericht über den 26. Kongress der Deutschen Gesellschaft für Psychologie in Tübingen 1968. Göttingen: Hogrefe. pp. 244–253.
  2. 1 2 Krauth, J.; Lienert, G. A. (1973). KFA. Die Konfigurationsfrequenzanalyse und ihre Anwendungen in Psychologie und Medizin[CFA. Configural frequency analysis and its application in psychology and medicine]. Freiburg: Alber.
  3. Holm, S. (1979). "A simple sequential rejective multiple test procedure". Scandinavian Journal of Statistics . 6 (2): 65–70. JSTOR   4615733.
  4. von Eye, A. (1990). Introduction to Configural Frequency Analysis: The search for types and antitypes in cross-classifications. Cambridge, UK: Cambridge University Press. ISBN   0521380901.
  5. Lautsch, E.; Weber, S. (1990). Konfigurationsfrequenzanalyse (KFA). Berlin: Volk und Wissen.
  6. Krauth, J. (1993). Einführung in die Konfigurationsfrequenzanalyse (KFA)[Introduction to Configural Frequency Analysis (CFA)]. Weinheim: Beltz, Psychologie Verlags Union. ISBN   3621271821.

Further reading