Grade: Grade 12 Subject: Mathematics Unit: Quantitative Reasoning SAT: ProblemSolving+DataAnalysis ACT: Math

Statistical Literacy

Learn

Why Statistical Literacy Matters

We encounter statistical claims daily - in news articles, advertisements, political debates, and scientific reports. Being statistically literate means being able to critically evaluate these claims, understand what the numbers actually mean, and recognize when statistics are being misused or misrepresented.

Statistical Literacy

Statistical literacy is the ability to understand, interpret, and critically evaluate statistical information in everyday life. It includes knowing how data is collected, what statistics can and cannot tell us, and recognizing common pitfalls in statistical reasoning.

Evaluating Statistical Claims

Key Questions to Ask

  1. Who collected the data? Are they unbiased? What's their motivation?
  2. How was data collected? Survey, experiment, observation?
  3. How large was the sample? Is it representative?
  4. What exactly was measured? How was it defined?
  5. What are the margins of error? Are they reported?
  6. Is correlation being confused with causation?
  7. Are comparisons fair? Are they comparing like with like?

Common Statistical Fallacies

1. Misleading Averages

The mean, median, and mode can give very different pictures of the same data. Reports often use whichever average serves their purpose.

Example: "Average home price is $500,000" - Is this the mean (inflated by mansions) or median (typical home)?

2. Base Rate Neglect

Ignoring how common something is in the population when evaluating probabilities.

Example: A test is 99% accurate. If a rare disease affects 1 in 10,000 people, a positive result is more likely to be a false positive than true positive!

3. Selection Bias

When the sample isn't representative of the population you're trying to study.

Example: Online polls attract people with strong opinions, missing moderate views.

4. Survivorship Bias

Focusing on successes while ignoring failures that are no longer visible.

Example: "This company started in a garage!" ignores the thousands of garage startups that failed.

Understanding Percentages and Ratios

What's StatedWhat It Really MeansWatch Out For
50% increase1.5 times the originalSmall base numbers exaggerate
50% decreaseHalf the originalMay still be significant
"Doubled"100% increaseFrom 1 to 2 is "doubling"
"Relative risk"Ratio compared to baselineAbsolute numbers may be small
Percentage pointsAbsolute difference in %Different from % change

Percentage Points vs. Percent Change

If support goes from 20% to 30%:

  • Percentage points: Increase of 10 percentage points
  • Percent change: 50% increase ((30-20)/20 = 0.5)

Both are correct, but they tell different stories!

Reading Graphs Critically

Common Graph Manipulation Techniques

  • Truncated Y-axis: Starting at a value other than zero exaggerates differences
  • Inappropriate scale: Stretching or compressing axes distorts patterns
  • Cherry-picked time periods: Choosing specific start/end dates to support a narrative
  • 3D effects: Makes accurate comparison of values difficult
  • Missing labels: Without units or context, values are meaningless

Sample Size and Margin of Error

Understanding Margins of Error

The margin of error indicates the range within which the true population value likely falls.

Margin of Error for proportions is approximately 1/sqrt(n)

A poll of 1,000 people typically has a margin of error around plus or minus 3%.

Smaller samples = larger margins of error = less precision

Examples

Example 1: Evaluating a Claim

Claim: "Our weight loss program works! Participants lost an average of 15 pounds."

Critical Analysis:

  • Sample size? Not mentioned - could be 5 people or 500
  • Selection bias? Did they only count people who completed the program?
  • Time frame? How long did the study last?
  • Control group? Without comparison, we don't know if results are unusual.
  • Mean vs. median? A few extreme losers could inflate the average.

Conclusion: The claim lacks crucial details needed to evaluate its validity.

Example 2: Base Rate Problem

Problem: A disease test is 99% accurate. The disease affects 1 in 1,000 people. If you test positive, what's the probability you have the disease?

Solution:

Imagine 100,000 people:

  • 100 actually have the disease
  • 99 of them test positive (99% sensitivity)
  • 99,900 don't have the disease
  • 999 of them test positive anyway (1% false positive rate)

Total positive tests: 99 + 999 = 1,098

True positives: 99

P(disease | positive test) = 99/1,098 = about 9%

Despite a "99% accurate" test, a positive result only means a 9% chance of having the disease!

Example 3: Misleading Graph

Situation: A graph shows Company A's stock rising from $100 to $102, but the Y-axis only shows $99 to $103.

Analysis:

Actual change: $102 - $100 = $2, which is 2%

The truncated Y-axis makes this modest gain appear dramatic.

How to spot this:

  • Check where the Y-axis starts - is it at zero?
  • Calculate the actual percentage change
  • Consider: Would this look significant on a full scale?

Fair representation: Y-axis from $0 would show this as a nearly flat line.

Example 4: Relative vs. Absolute Risk

Headline: "New drug reduces heart attack risk by 50%!"

Digging Deeper:

Suppose in the study:

  • Control group: 2 out of 1,000 had heart attacks (0.2%)
  • Drug group: 1 out of 1,000 had heart attacks (0.1%)

Relative risk reduction: (2-1)/2 = 50% - technically true!

Absolute risk reduction: 0.2% - 0.1% = 0.1 percentage points

Number needed to treat: 1,000 people need the drug to prevent 1 heart attack

Conclusion: The 50% sounds impressive, but the actual benefit is small.

Example 5: Confounding Variables

Observation: Countries with more Nobel Prize winners also have higher chocolate consumption. Does chocolate make you smarter?

Analysis:

This is almost certainly NOT causation. Possible confounders:

  • Wealth: Richer countries have more resources for education AND chocolate
  • Education systems: Better education leads to both research success and luxury goods consumption

Conclusion: There's likely a confounding variable (national wealth) that explains both.

Practice

Test your ability to critically evaluate statistical claims.

1. "9 out of 10 dentists recommend our toothpaste." What question should you ask?

A) What brand do the dentists use?   B) How were dentists selected?   C) How much does it cost?   D) What flavor is it?

2. A poll with margin of error plus or minus 4% shows Candidate A at 52% and B at 48%. What can we conclude?

A) A will definitely win   B) Race is too close to call   C) B is likely to win   D) Poll is invalid

3. "Crime doubled from 2 incidents to 4." This is an example of:

A) Good statistical reporting   B) Misleading use of small numbers   C) Survivorship bias   D) Selection bias

4. To establish causation, you need:

A) Strong correlation   B) A controlled experiment   C) Large sample size   D) Multiple surveys

5. Average CEO pay is $15 million. Average worker pay is $50,000. The ratio is often criticized because:

A) CEOs work harder   B) A few extreme CEO salaries skew the mean   C) Workers are undercounted   D) The comparison is irrelevant

6. An online poll asks "Do you support Policy X?" This likely suffers from:

A) Random sampling   B) Self-selection bias   C) Too large a sample   D) Measurement error

7. Unemployment "dropped" from 5% to 4.8%. This is a change of:

A) 0.2%   B) 0.2 percentage points   C) 4%   D) Both A and B

8. "People who eat breakfast earn more money." This could be explained by:

A) Breakfast causes wealth   B) Wealthy people eat more breakfast   C) A third factor like stable lifestyle   D) B or C

9. A graph shows dramatic growth, but the Y-axis starts at 95%. This technique is called:

A) Data visualization   B) Truncated axis   C) Statistical significance   D) Normalization

10. A study of successful entrepreneurs finds most dropped out of college. This ignores:

A) Their original majors   B) Dropouts who failed (survivorship bias)   C) Their height   D) Their location

Click to reveal answers
  1. B) How were dentists selected? - Selection method determines if sample is representative
  2. B) Race is too close to call - The intervals overlap
  3. B) Misleading use of small numbers - Doubling sounds dramatic but is just 2 more incidents
  4. B) A controlled experiment - Only way to isolate cause and effect
  5. B) A few extreme CEO salaries skew the mean - Median would be more representative
  6. B) Self-selection bias - Only motivated people respond to online polls
  7. B) 0.2 percentage points - "Percent" would be relative change (4%)
  8. D) B or C - Could be reverse causation or confounding
  9. B) Truncated axis - Exaggerates small changes
  10. B) Dropouts who failed (survivorship bias) - We only see the successes

Check Your Understanding

1. Why is it important to know who funded a study?

Show answer

Funding sources can indicate potential bias. Studies funded by industries may have incentives to find favorable results. This doesn't mean the research is wrong, but it warrants extra scrutiny. Look for: independent replication, peer review, disclosure of conflicts of interest, and whether negative results are published.

2. Explain why a graph with a truncated Y-axis can be misleading.

Show answer

When the Y-axis doesn't start at zero, small differences appear much larger than they are. For example, a stock going from $100 to $102 (2% change) looks like it nearly tripled if the axis runs from $99 to $103. Our visual perception responds to the size of the bars/lines, not the actual numbers. Always check axis labels and calculate actual percentage changes.

3. What's the difference between statistical significance and practical significance?

Show answer

Statistical significance means the result is unlikely due to chance. Practical significance means the result matters in the real world. With large samples, tiny effects become statistically significant. A drug might reduce blood pressure by 0.5 mmHg with very low p-value (statistically significant) but this effect is too small to matter clinically (not practically significant). Always consider effect size and real-world impact.

4. How can you protect yourself from being misled by statistics?

Show answer

Strategies include: (1) Ask about sample size and selection method, (2) Look for the actual numbers, not just relative changes, (3) Check if causation is being claimed from correlation, (4) Look for conflicts of interest, (5) Seek multiple sources, (6) Be suspicious of dramatic claims, (7) Check graph axes and scales, (8) Ask what's missing or not being reported, (9) Consider base rates and context, (10) Distinguish between percentage and percentage points.

🚀 Next Steps

  • Review any concepts that felt challenging
  • Move on to the next lesson when ready
  • Return to practice problems periodically for review