Statistics and Probability for 11th Graders

Joshua Marie

2025-11-20

What is Statistics?

Definition

Statistics is the science of collecting, organizing, analyzing, interpreting, and presenting data. Used in research, business, science, and everyday decision-making

Two kinds of statistics:

  • Descriptive Statistics: The way to summarize and describe data (mean, median, mode, graphs)
  • Inferential Statistics: Meticulous way to make generalizations, i.e. predictions or inferences, about a population based on sample data

What is Probability?

Definition

Probability is the measure of the likelihood that an event will occur.

  • Expressed as a decimal number between 0 and 1 (or 0% to 100%)
  • \(P(E) = 0\) means the event is impossible
  • \(P(E) = 1\) means the event is certain
  • Helps us make predictions about random events

Calculating Basic Probability

Formula for Probability

\[P(\text{Event}) = \frac{\text{Number of favorable outcomes}}{\text{Total number of possible outcomes}}=\frac{x}{n}\]

Example: Rolling a die

What is the probability of rolling a 4 in a die?

  • Favorable outcomes: {4} \(\rightarrow\) 1 outcome
  • Total outcomes: \(\{1,\ 2,\ 3,\ 4,\ 5,\ 6\}\) \(\rightarrow\) 6 outcomes
  • \(P(\text{rolling a 4}) = \frac{1}{6}\)

Sample Space

Definition

A sample space (S) is the collection or set of all possible outcomes of a random experiment.

  • Represented using the symbol “S” or \(\Omega\) in some textbooks
  • Contains all outcomes that could occur
  • Can be finite or infinite

Examples:

  • Tossing a coin: \(S = \{H, T\}\)
  • Rolling a die: \(S = \{1,\ 2,\ 3,\ 4,\ 5,\ 6\}\)
  • Tossing two coins: \(S = \{HH,\ HT,\ TH,\ TT\}\)

Events

Definition

An event is a subset of possible outcomes from the sample space.

  • Can be a single outcome or multiple outcomes
  • Denoted by capital letters (A, B, C)

Important

When you calculate the probability of the event \(P(E)\), it is ALWAYS a decimal number, a fractional number, ranges from 0 to 1.

For example:

When tossing a coin, what’s the probability of getting heads? There should be two outcomes: getting a head & getting a tail . The sample space would be: \(S=\{H,\ T\}\) And the event A is getting head Thus, \(x=1;\ n=2\), and the probability would be:

\[P(A)=\frac{1}{2}=0.5\]

Events

Definition

  • An event is any subset of possible outcomes in the sample space \(\Omega\).

  • Compound event: contains two or more outcomes (e.g., “rolling an even number”)

Calculating Probability of Events

Practice Problem:

A bag contains 5 red balls, 3 blue balls, and 2 green balls.

  1. What is the probability of drawing a red ball?
  2. What is the probability of drawing a blue or green ball?
  3. What is the probability of NOT drawing a red ball?
  1. \(P(\text{red}) = \frac{5}{10} = \frac{1}{2}\)
  2. \(P(\text{blue or green}) = \frac{3}{10} + \frac{2}{10} = \frac{5}{10} = \frac{1}{2}\)
  3. \(P(\text{not red}) = 1 - P(\text{red}) = 1 - \frac{1}{2} = \frac{1}{2}\)

Events

Mutually Exclusive (Disjoint) Events

Two events \(A\) and \(B\) are mutually exclusive if they have no outcomes in common:

\[A \cap B = \emptyset\]

Consequence for probabilities:

\[P(A \cup B) = P(A) + P(B)\] (no overlap to double-count)

Classic example – Die roll

  • \(A\): Rolling an even number \(\{2,\ 4,\ 6\}\)
  • \(B\): Rolling an odd number \(\{1,\ 3,\ 5\}\)

\(\rightarrow\) \(A \cap B = \emptyset\) \(\rightarrow\) mutually exclusive

Another example: Drawing one card

  • \(A\): Drawing a heart
  • \(B\): Drawing a spade

\(\rightarrow\) also mutually exclusive

Events

Non-Mutually Exclusive Events

Two events \(A\) and \(B\) are non-mutually exclusive if they share at least one outcome:

\[A \cap B \neq \emptyset\]

Addition rule becomes:

\[P(A \cup B) = P(A) + P(B) - P(A \cap B)\] (we subtract the overlap so it’s not counted twice)

Die roll example

  • \(A\): Rolling an even number: \(\{2,\ 4,\ 6\}\)
  • \(B\): Rolling a number $ >4 : {5, 6}$

\(\rightarrow\) \(A \cap B = \{6\} \neq \emptyset\) \(\rightarrow\) not mutually exclusive

\(P(A \cup B) = P(A) + P(B) - P(A \cap B) = \frac{3}{6} + \frac{2}{6} - \frac{1}{6} = \frac{4}{6} = \frac{2}{3}\)

Random Variables

Definition

A random variable is a function that assigns a numerical value to each outcome in the sample space of a random experiment.

Notation: We use capital letters (X, Y, Z) for random variables

How it works:

  1. Perform a random experiment \(\rightarrow\) get outcomes in sample space \(S\)
  2. The random variable \(X\) maps each outcome to a number
  3. We can then calculate probabilities: \(P(X = x)\)

Intuition: Tossing two coins

  • Sample space: \(S = \{HH,\ HT,\ TH,\ TT\}\)
  • Let \(X\) = number of heads
  • Mapping: \(HH \rightarrow 2\), \(HT \rightarrow 1\), \(TH \rightarrow 1\), \(TT \rightarrow 0\)
  • So \(X\) can be 0, 1, or 2
  • \(P(X = 1) = \frac{2}{4} = \frac{1}{2}\) (since HT and TH both give 1 head)

Random Variables

Definition

A random variable is a function that assigns numerical values to outcomes of a random experiment.

  • Maps outcomes to real numbers
  • Denoted by capital letters (X, Y, Z)
  • Can be discrete or continuous

Example

Rolling a die: \(X = \text{number shown}\)

  • \(X\) can be 1, 2, 3, 4, 5, or 6
  • Each outcome has equal probability
  • \(P(X = k) = \frac{1}{6}; k \in {\{1,2,3,4,5,6}\}\)

Types of Random Variables

Discrete Random Variables

Takes on countable values

Key Property: Can list all discrete possible values

Examples:

  • Number of heads in 10 coin flips
  • Number of students absent
  • Number of cars passing by
  • Number of balls drawn in an urn

Continuous Random Variables

Takes on uncountably infinite values

Key Property: Any value in an interval is possible

Examples:

  • Height of students
  • Time until next bus arrives
  • Temperature at noon
  • Waiting time

Probability Distribution

For Discrete Random Variables

A probability distribution assigns probabilities to each possible value.

Requirements:

  • \(0 \leq P(X = x) \leq 1\) for all \(x\)
  • \(\sum_{x}^{\infty}P(X = x) = 1\) (sum over all possible values)

Example: Rolling a die

\(X\) 1 2 3 4 5 6
\(P(X)\) 1/6 1/6 1/6 1/6 1/6 1/6

Expected Value (Mean)

The expected value E(X) is the average value we expect from a random variable.

Formula

\[E(X) = \sum_{i}^{\infty} x_i \cdot P(X = x_i)\]

Example: Rolling a die

\[E(X) = 1 \cdot \frac{1}{6} + 2 \cdot \frac{1}{6} + 3 \cdot \frac{1}{6} + 4 \cdot \frac{1}{6} + 5 \cdot \frac{1}{6} + 6 \cdot \frac{1}{6}\]

\[E(X) = \frac{21}{6} = 3.5\]

Variance and Standard Deviation

Variance \(\sigma^2\)

Measures how spread out the values are from the mean.

\[\text{Var}(X) = E[(X - \mu)^2]\] \[= \sum_{i} (x_i - \mu)^2 \cdot P(X = x_i)\]

where \(\mu = E(X)\)

Variance and Standard Deviation

Variance \(\sigma^2\)

Measures how spread out the values are from the mean.

\[\text{Var}(X) = E[(X - \mu)^2]\] \[= \sum_{i} (x_i - \mu)^2 \cdot P(X = x_i)\]

where \(\mu = E(X)\)

Standard Deviation \(\sigma\)

The square root of variance (same units as X).

\[\sigma = \sqrt{\text{Var}(X)}\]

Interpretation:

  • Small \(\sigma\): values clustered near mean
  • Large \(\sigma\): values spread out

Where Do All These Distributions Come From?

Many real-life situations follow repeatable patterns. When we study those patterns carefully, we discover a handful of probability distributions that keep showing up again and again.

The normal distribution (the bell curve) is the most famous distribution out there. Yes, it is only a member of a much larger family.

Common Discrete Distributions

Here are common discrete probability distribution.

Scenario: A single trial with only two possible outcomes (success or failure)

\[P(X = x) = \begin{cases} p & \text{if } x = 1 \text{ (success)} \\ 1-p & \text{if } x = 0 \text{ (failure)} \end{cases}\]

\[P(X = x)=p^x(1-p)^{1-x}\]

Properties: - \(E(X) = p\) - \(\text{Var}(X) = p(1-p)\)

Example: Flipping a coin once, \(X = 1\) if heads, \(X = 0\) if tails - \(p = 0.5\) - \(E(X) = 0.5\) - \(\text{Var}(X) = 0.5(0.5) = 0.25\)

Scenario: n independent trials, each with probability p of success

\[P(X = k) = \binom{n}{k} p^k (1-p)^{n-k}\]

Example: Flipping a coin 10 times, X = number of heads

  • \(n = 10\), \(p = 0.5\)
  • \(E(X) = np = 5\)
  • \(\text{Var}(X) = np(1-p) = 2.5\)

Practice Problem

Problem: A bag contains 3 red balls and 2 blue balls. You draw 2 balls without replacement. Let X = number of red balls drawn.

Find:

  1. The probability distribution of \(X\)
  2. \(E(X)\)
  3. \(Var(X)\)

Solution

Probability Distribution:

  • \(P(X = 0) = P(\text{2 blue}) = (2/5)(1/4) = 2/20 = 1/10\)
  • \(P(X = 1) = P(\text{1 red, 1 blue}) = (2)(3/5)(2/4) = 12/20 = 3/5\)
  • \(P(X = 2) = P(\text{2 red}) = (3/5)(2/4) = 6/20 = 3/10\)

Expected Value:

E(X) = 0(1/10) + 1(3/5) + 2(3/10) = 0 + 3/5 + 6/10 = 12/10 = 1.2

Variance:

Var(X) = (0-1.2)²(1/10) + (1-1.2)²(3/5) + (2-1.2)²(3/10) = 0.36