Sample space

Last updated

In probability theory, the sample space (also called sample description space, [1] possibility space, [2] or outcome space [3] ) of an experiment or random trial is the set of all possible outcomes or results of that experiment. [4] A sample space is usually denoted using set notation, and the possible ordered outcomes, or sample points, [5] are listed as elements in the set. It is common to refer to a sample space by the labels S, Ω, or U (for "universal set"). The elements of a sample space may be numbers, words, letters, or symbols. They can also be finite, countably infinite, or uncountably infinite. [6]

Contents

A subset of the sample space is an event, denoted by . If the outcome of an experiment is included in , then event has occurred. [7]

For example, if the experiment is tossing a single coin, the sample space is the set , where the outcome means that the coin is heads and the outcome means that the coin is tails. [8] The possible events are , , , and . For tossing two coins, the sample space is , where the outcome is if both coins are heads, if the first coin is heads and the second is tails, if the first coin is tails and the second is heads, and if both coins are tails. [9] The event that at least one of the coins is heads is given by .

For tossing a single six-sided die one time, where the result of interest is the number of pips facing up, the sample space is . [10]

A well-defined, non-empty sample space is one of three components in a probabilistic model (a probability space). The other two basic elements are: a well-defined set of possible events (an event space), which is typically the power set of if is discrete or a σ-algebra on if it is continuous, and a probability assigned to each event (a probability measure function). [11]

A visual representation of a finite sample space and events. The red oval is the event that a number is odd, and the blue oval is the event that a number is prime. Sample space.png
A visual representation of a finite sample space and events. The red oval is the event that a number is odd, and the blue oval is the event that a number is prime.

A sample space can be represented visually by a rectangle, with the outcomes of the sample space denoted by points within the rectangle. The events may be represented by ovals, where the points enclosed within the oval make up the event. [12]

Conditions of a sample space

A set with outcomes (i.e. ) must meet some conditions in order to be a sample space: [13]

For instance, in the trial of tossing a coin, one possible sample space is , where is the outcome where the coin lands heads and is for tails. Another possible sample space could be . Here, denotes a rainy day and is a day where it is not raining. For most experiments, would be a better choice than , as an experimenter likely does not care about how the weather affects the coin toss.

Multiple sample spaces

For many experiments, there may be more than one plausible sample space available, depending on what result is of interest to the experimenter. For example, when drawing a card from a standard deck of fifty-two playing cards, one possibility for the sample space could be the various ranks (Ace through King), while another could be the suits (clubs, diamonds, hearts, or spades). [4] [14] A more complete description of outcomes, however, could specify both the denomination and the suit, and a sample space describing each individual card can be constructed as the Cartesian product of the two sample spaces noted above (this space would contain fifty-two equally likely outcomes). Still other sample spaces are possible, such as right-side up or upside down, if some cards have been flipped when shuffling.

Equally likely outcomes

Flipping a coin leads to a sample space composed of two outcomes that are almost equally likely. Coin tossing.JPG
Flipping a coin leads to a sample space composed of two outcomes that are almost equally likely.
Up or down? Flipping a brass tack leads to a sample space composed of two outcomes that are not equally likely. Brass thumbtack.jpg
Up or down? Flipping a brass tack leads to a sample space composed of two outcomes that are not equally likely.

Some treatments of probability assume that the various outcomes of an experiment are always defined so as to be equally likely. [15] For any sample space with equally likely outcomes, each outcome is assigned the probability . [16] However, there are experiments that are not easily described by a sample space of equally likely outcomes—for example, if one were to toss a thumb tack many times and observe whether it landed with its point upward or downward, there is no physical symmetry to suggest that the two outcomes should be equally likely. [17]

Though most random phenomena do not have equally likely outcomes, it can be helpful to define a sample space in such a way that outcomes are at least approximately equally likely, since this condition significantly simplifies the computation of probabilities for events within the sample space. If each individual outcome occurs with the same probability, then the probability of any event becomes simply: [18] :346–347

For example, if two fair six-sided dice are thrown to generate two uniformly distributed integers, and , each in the range from 1 to 6, inclusive, the 36 possible ordered pairs of outcomes constitute a sample space of equally likely events. In this case, the above formula applies, such as calculating the probability of a particular sum of the two rolls in an outcome. The probability of the event that the sum is five is , since four of the thirty-six equally likely pairs of outcomes sum to five.

If the sample space was all of the possible sums obtained from rolling two six-sided dice, the above formula can still be applied because the dice rolls are fair, but the number of outcomes in a given event will vary. A sum of two can occur with the outcome , so the probability is . For a sum of seven, the outcomes in the event are , so the probability is . [19]

Simple random sample

In statistics, inferences are made about characteristics of a population by studying a sample of that population's individuals. In order to arrive at a sample that presents an unbiased estimate of the true characteristics of the population, statisticians often seek to study a simple random sample—that is, a sample in which every individual in the population is equally likely to be included. [18] :274–275 The result of this is that every possible combination of individuals who could be chosen for the sample has an equal chance to be the sample that is selected (that is, the space of simple random samples of a given size from a given population is composed of equally likely outcomes). [20]

Infinitely large sample spaces

In an elementary approach to probability, any subset of the sample space is usually called an event. [9] However, this gives rise to problems when the sample space is continuous, so that a more precise definition of an event is necessary. Under this definition only measurable subsets of the sample space, constituting a σ-algebra over the sample space itself, are considered events.

An example of an infinitely large sample space is measuring the lifetime of a light bulb. The corresponding sample space would be [0, ∞). [9]

See also

Related Research Articles

<span class="mw-page-title-main">Entropy (information theory)</span> Expected amount of information needed to specify the output of a stochastic data source

In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable , which takes values in the alphabet and is distributed according to :

<span class="mw-page-title-main">Elementary event</span>

In probability theory, an elementary event, also called an atomic event or sample point, is an event which contains only a single outcome in the sample space. Using set theory terminology, an elementary event is a singleton. Elementary events and their corresponding outcomes are often written interchangeably for simplicity, as such an event corresponding to precisely one outcome.

<span class="mw-page-title-main">Event (probability theory)</span> In statistics and probability theory, set of outcomes to which a probability is assigned

In probability theory, an event is a set of outcomes of an experiment to which a probability is assigned. A single outcome may be an element of many different events, and different events in an experiment are usually not equally likely, since they may include very different groups of outcomes. An event consisting of only a single outcome is called an elementary event or an atomic event; that is, it is a singleton set. An event that has more than one possible outcomes is called compound event. An event is said to occur if contains the outcome of the experiment. The probability that an event occurs is the probability that contains the outcome of an experiment. An event defines a complementary event, namely the complementary set, and together these define a Bernoulli trial: did the event occur or not?

<span class="mw-page-title-main">Probability theory</span> Branch of mathematics concerning probability

Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set of axioms. Typically these axioms formalise probability in terms of a probability space, which assigns a measure taking values between 0 and 1, termed the probability measure, to a set of outcomes called the sample space. Any specified subset of the sample space is called an event.

<span class="mw-page-title-main">Probability distribution</span> Mathematical function for the probability a given outcome occurs in an experiment

In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon in terms of its sample space and the probabilities of events.

<span class="mw-page-title-main">Random variable</span> Variable representing a random phenomenon

A random variable is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' can be misleading as its mathematical definition is not actually random nor a variable, but rather it is a function from possible outcomes in a sample space to a measurable space, often to the real numbers.

<span class="mw-page-title-main">Probability space</span> Mathematical concept

In probability theory, a probability space or a probability triple is a mathematical construct that provides a formal model of a random process or "experiment". For example, one can define a probability space which models the throwing of a die.

<span class="mw-page-title-main">Bernoulli process</span> Random process of binary (boolean) random variables

In probability and statistics, a Bernoulli process is a finite or infinite sequence of binary random variables, so it is a discrete-time stochastic process that takes only two values, canonically 0 and 1. The component Bernoulli variablesXi are identically distributed and independent. Prosaically, a Bernoulli process is a repeated coin flipping, possibly with an unfair coin. Every variable Xi in the sequence is associated with a Bernoulli trial or experiment. They all have the same Bernoulli distribution. Much of what can be said about the Bernoulli process can also be generalized to more than two outcomes ; this generalization is known as the Bernoulli scheme.

<span class="mw-page-title-main">Bernoulli trial</span> Any experiment with two possible random outcomes

In the theory of probability and statistics, a Bernoulli trial is a random experiment with exactly two possible outcomes, "success" and "failure", in which the probability of success is the same every time the experiment is conducted. It is named after Jacob Bernoulli, a 17th-century Swiss mathematician, who analyzed them in his Ars Conjectandi (1713).

<span class="mw-page-title-main">Probability mass function</span> Discrete-variable probability distribution

In probability and statistics, a probability mass function is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes it is also known as the discrete probability density function. The probability mass function is often the primary means of defining a discrete probability distribution, and such functions exist for either scalar or multivariate random variables whose domain is discrete.

<span class="mw-page-title-main">Martingale (probability theory)</span> Model in probability theory

In probability theory, a martingale is a sequence of random variables for which, at a particular time, the conditional expectation of the next value in the sequence is equal to the present value, regardless of all prior values.

In probability theory, an event is said to happen almost surely if it happens with probability 1. In other words, the set of possible exceptions may be non-empty, but it has probability 0. The concept is analogous to the concept of "almost everywhere" in measure theory. In probability experiments on a finite sample space with a non-zero probability for each outcome, there is no difference between almost surely and surely ; however, this distinction becomes important when the sample space is an infinite set, because an infinite set can have non-empty subsets of probability 0.

In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable. It can be thought of as an alternative way of expressing probability, much like odds or log-odds, but which has particular mathematical advantages in the setting of information theory.

In statistics, the question of checking whether a coin is fair is one whose importance lies, firstly, in providing a simple problem on which to illustrate basic ideas of statistical inference and, secondly, in providing a simple problem that can be used to compare various competing methods of statistical inference, including decision theory. The practical problem of checking whether a coin is fair might be considered as easily solved by performing a sufficiently large number of trials, but statistics and probability theory can provide guidance on two types of question; specifically those of how many trials to undertake and of the accuracy of an estimate of the probability of turning up heads, derived from a given sample of trials.

In probability theory, random element is a generalization of the concept of random variable to more complicated spaces than the simple real line. The concept was introduced by Maurice Fréchet (1948) who commented that the “development of probability theory and expansion of area of its applications have led to necessity to pass from schemes where (random) outcomes of experiments can be described by number or a finite set of numbers, to schemes where outcomes of experiments represent, for example, vectors, functions, processes, fields, series, transformations, and also sets or collections of sets.”

<span class="mw-page-title-main">Fair coin</span>

In probability theory and statistics, a sequence of independent Bernoulli trials with probability 1/2 of success on each trial is metaphorically called a fair coin. One for which the probability is not 1/2 is called a biased or unfair coin. In theoretical studies, the assumption that a coin is fair is often made by referring to an ideal coin.

<span class="mw-page-title-main">Experiment (probability theory)</span> Procedure that can be infinitely repeated, with a well-defined set of outcomes

In probability theory, an experiment or trial is any procedure that can be infinitely repeated and has a well-defined set of possible outcomes, known as the sample space. An experiment is said to be random if it has more than one possible outcome, and deterministic if it has only one. A random experiment that has exactly two possible outcomes is known as a Bernoulli trial.

<span class="mw-page-title-main">Conditional probability</span> Probability of an event occurring, given that another event has already occurred

In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption, assertion or evidence) is already known to have occurred. This particular method relies on event B occurring with some sort of relationship with another event A. In this event, the event B can be analyzed by a conditional probability with respect to A. If the event of interest is A and the event B is known or assumed to have occurred, "the conditional probability of A given B", or "the probability of A under the condition B", is usually written as P(A|B) or occasionally PB(A). This can also be understood as the fraction of probability B that intersects with A, or the ratio of the probabilities of both events happening to the "given" one happening (how many times A occurs rather than not assuming B has occurred): .

<span class="mw-page-title-main">Outcome (probability)</span> Possible result of an experiment or trial

In probability theory, an outcome is a possible result of an experiment or trial. Each possible outcome of a particular experiment is unique, and different outcomes are mutually exclusive. All of the possible outcomes of an experiment form the elements of a sample space.

In probability theory, the van den Berg–Kesten (BK) inequality or van den Berg–Kesten–Reimer (BKR) inequality states that the probability for two random events to both happen, and at the same time one can find "disjoint certificates" to show that they both happen, is at most the product of their individual probabilities. The special case for two monotone events was first proved by van den Berg and Kesten in 1985, who also conjectured that the inequality holds in general, not requiring monotonicity. Reimer later proved this conjecture. The inequality is applied to probability spaces with a product structure, such as in percolation problems.

References

  1. Stark, Henry; Woods, John W. (2002). Probability and Random Processes with Applications to Signal Processing (3rd ed.). Pearson. p. 7. ISBN   9788177583564.
  2. Forbes, Catherine; Evans, Merran; Hastings, Nicholas; Peacock, Brian (2011). Statistical Distributions (4th ed.). Wiley. p.  3. ISBN   9780470390634.
  3. Hogg, Robert; Tannis, Elliot; Zimmerman, Dale (December 24, 2013). Probability and Statistical Inference. Pearson Education, Inc. p. 10. ISBN   978-0321923271. The collection of all possible outcomes... is called the outcome space.
  4. 1 2 Albert, Jim (1998-01-21). "Listing All Possible Outcomes (The Sample Space)". Bowling Green State University. Retrieved 2013-06-25.
  5. Soong, T. T. (2004). Fundamentals of probability and statistics for engineers. Chichester: Wiley. ISBN   0-470-86815-5. OCLC   55135988.
  6. 1 2 3 "UOR_2.1". web.mit.edu. Retrieved 2019-11-21.
  7. Ross, Sheldon (2010). A First Course in Probability (PDF) (8th ed.). Pearson Prentice Hall. p. 23. ISBN   978-0136033134.
  8. Dekking, F.M. (Frederik Michel), 1946- (2005). A modern introduction to probability and statistics : understanding why and how. Springer. ISBN   1-85233-896-2. OCLC   783259968.{{cite book}}: CS1 maint: multiple names: authors list (link)
  9. 1 2 3 "Sample Space, Events and Probability" (PDF). Mathematics at Illinois.
  10. Larsen, R. J.; Marx, M. L. (2001). An Introduction to Mathematical Statistics and Its Applications (3rd ed.). Upper Saddle River, NJ: Prentice Hall. p. 22. ISBN   9780139223037.
  11. LaValle, Steven M. (2006). Planning Algorithms (PDF). Cambridge University Press. p. 442.
  12. "Sample Spaces, Events, and Their Probabilities". saylordotorg.github.io. Retrieved 2019-11-21.
  13. Tsitsiklis, John (Spring 2018). "Sample Spaces". Massachusetts Institute of Technology. Retrieved July 9, 2018.
  14. Jones, James (1996). "Stats: Introduction to Probability - Sample Spaces". Richland Community College. Retrieved 2013-11-30.
  15. Foerster, Paul A. (2006). Algebra and Trigonometry: Functions and Applications, Teacher's Edition (Classics ed.). Prentice Hall. p.  633. ISBN   0-13-165711-9.
  16. "Equally Likely outcomes" (PDF). University of Notre Dame.
  17. "Chapter 3: Probability" (PDF). Coconino Community College.
  18. 1 2 Yates, Daniel S.; Moore, David S.; Starnes, Daren S. (2003). The Practice of Statistics (2nd ed.). New York: Freeman. ISBN   978-0-7167-4773-4. Archived from the original on 2005-02-09.
  19. "Probability: Rolling Two Dice". www.math.hawaii.edu. Retrieved 2021-12-17.
  20. "Simple Random Samples". web.ma.utexas.edu. Retrieved 2019-11-21.