📊 Statistics 1: The Master Encyclopedia (V5)

Welcome to the ultimate guide for Statistics 1. This document is designed to be the “goated” repository for every concept, formula, and nuance in the course.


📔 Volume I: The Genesis of Data & Descriptive Foundations


🔬 Week 1: The Anatomy of Statistical Inquiry

1.1 The Epistemology of Statistics

Statistics is the mathematical framework for decision-making under uncertainty. It translates raw observation into actionable intelligence.

  • Descriptive Statistics: Summarizing and visualizing the data we have on hand.
  • Inferential Statistics: Making predictions or generalizations about a Population based on a Sample.

Key Definitions:

  • Population (): The entire collection of interest.
  • Sample (): The observed subset.
  • Parameter: A characteristic of the population (e.g., Population Mean ).
  • Statistic: A characteristic of the sample (e.g., Sample Mean ).

1.2 The Taxonomy of Variables

TypeSub-typeDefinitionExample
CategoricalNominalIdentity only, no orderBlood type, Gender
OrdinalMeaningful order, distance unknownRanks, Likert scale
NumericalDiscreteCountable gaps (integers)No. of children
ContinuousInfinite values in a rangeHeight, Weight

1.3 The Scale Hierarchy (S.S. Stevens)

  1. Nominal: Identity ().
  2. Ordinal: Ranking ().
  3. Interval: Meaningful differences (). No “True Zero” (e.g., ).
  4. Ratio: True Zero exists (). Zero means absence of property (e.g., kg).

CAUTION

The “True Zero” Trap: Always ask: “If this is 0, does it mean the thing doesn’t exist?” If yes, it’s Ratio. If no, it’s likely Interval.



🎨 Week 2: Graphical Representations & Frequency Geometry

2.1 Frequency Distributions

Before visualizing, we must structure raw data into Frequency Tables.

  • Relative Frequency: (Proportion of the total).
  • Percent Frequency: Relative Frequency .

2.2 Visualizing Categorical Data

The goal is to show the distribution of observations across categories.

1. The Bar Chart

  • Geometry: Rectangular bars where .
  • Best Use: Comparing specific values across categories.
  • Variation: Pareto Chart. A bar chart sorted in descending order of frequency ( rule application).

2. The Pie Chart

  • Geometry: Circular sectors where .
  • Formula: Sector Angle .
  • Best Use: Highlighting the “part-to-whole” relationship.

2.3 The “Golden Rule” of Visuals: The Area Principle

The visual area of a graphical element MUST be proportional to the value it represents.

  • Violation: 3D effects, varying thicknesses, or uneven widths in bar charts. These are designed to mislead and are “Cardinal Sins” in Statistics.

2.4 Case Study: Placement Analysis

  • Sector: Software (200), Core (150), Analytics (100), Other (50).
  • Total (): 500.
  • Software RF: .
  • Software Angle: .
  • Pareto Logic: Software would be the first bar, followed by Core, then Analytics.

📐 Volume II: Measures of Centrality & Linear Relationships

📏 Week 3: The Mechanics of Centrality & Dispersion

Numerical data requires summary metrics that capture where the data “centers” and how much it “leaks” (spreads).

3.1 Measures of Central Tendency

  1. Arithmetic Mean ():
    • .
    • Philosophical Insight: The mean is the “Balance Point” of the data. If you placed weights on a beam at the data points, the mean is where it would balance perfectly.
    • Sensitivity: Extremely sensitive to outliers. One massive value pulls the mean toward it.
  2. Median ():
    • The middle value of a sorted dataset.
    • If is odd: .
    • If is even: Average of values at ranks and .
    • Robustness: The median is a “Robust Statistic.” It ignores outliers.
  3. Mode:
    • The most frequent observation.
    • The only measure valid for Nominal data.

3.2 Measures of Dispersion (Spread)

  1. Variance ( or ):
    • Average squared deviation from the mean.
    • Sample Variance: .
    • Why ? (Bessel’s Correction): Using would underestimate the true population variance. makes the sample variance an “Unbiased Estimator.”
  2. Standard Deviation ( or ):
    • . Brings the units back to the original scale.
  3. Coefficient of Variation (CV):
    • .
    • Use: Comparing spread across different units (e.g., Variability of height in cm vs weight in kg).

3.3 Percentiles & The Anatomy of a Box Plot

  • Percentiles (): Value such that of data is at or below it.
  • Quartiles: .
  • IQR (Interquartile Range): .
  • Outlier Detection (Tukey’s Fences):
    • Lower Fence:
    • Upper Fence:
    • Values outside these fences are statistically suspected outliers.

When we observe two variables for the same case, we search for Correlation—the degree to which they move together.

  • Positive Trend: As increases, tends to increase.
  • Negative Trend: As increases, tends to decrease.
  • No Trend: Points are scattered randomly with no discernible line.

4.2 Covariance & Pearson’s

  1. Covariance ():
    • Measures the direction of the relationship.
    • .
    • If positive, they move together. If negative, they move oppositely.
  2. Correlation Coefficient ():
    • Normalizes covariance to a range of .
    • .
    • Strength:
      • : Strong linear relationship.
      • : Moderate linear relationship.
      • : No linear relationship (they might still have a curved relationship!).

4.3 Linear Transformations & Correlation

This is a high-frequency exam topic. If we transform and :

  • Covariance Result: .
  • Correlation Result:
    • if and have the same sign.
    • if and have opposite signs.
  • Invariance: Adding/Subtracting constants () has zero effect on .

4.4 Scatter Plot Interpretation Strategy

  • Step 1: Identify the independent and dependent variables.
  • Step 2: Look for the overall “cloud” shape.
  • Step 3: Is it linear? (If it’s a parabola, might be 0 even if there’s a clear relationship).
  • Step 4: Outliers? A single outlier far from the line can tank a high correlation.

🎲 Volume III: The Logic of Chance (Combinatorics & Probability)

🔢 Week 5: Countable Assemblies (Combinatorics)

Counting is the backbone of discrete probability. We must master the art of determining the size of the Sample Space () and the Event Space ().

5.1 The Multiplication Principle (The “AND” Rule)

If task A can be done in ways and task B in ways, doing A AND B takes ways.

  • Example: 3 shirts and 4 pants outfits.

5.2 Permutations (Order Matters)

Use when the sequence is distinct (e.g., gold/silver/bronze medals).

  1. Basic Permutation: .
  2. Identical Items (The Anagram Rule): If there are items of one kind, of another, total ways to arrange items:
    • Example: “STATISTICS” (10 letters: 3 S, 3 T, 2 I). Ways .
  3. Circular Permutations: If items are in a circle, ways .

5.3 Combinations (Order Doesn’t Matter)

Use when only the selection counts (e.g., a committee of 3).

  • Formula: .
  • Symmetry: (Choosing to take is the same as choosing to leave).

5.4 The “Bundle & Gap” Strategy

  • Bundle Method: When items must be together. Bundle them as 1 item, arrange the bundle internal elements, then arrange the new set.
  • Gap Method: When items must NOT be together. Arrange the other items first, then place the restricted items in the “gaps” between them.

🌩️ Weeks 6-8: The Calculus of Uncertainty (Probability)

6.1 The Axiomatic Foundation

Probability is a measure of belief or frequency assigned to an event .

  • Axiom 1: .
  • Axiom 2: .
  • Axiom 3: If and are Mutually Exclusive (), then .

6.2 The “Independence” vs. “disjoint” Trap

  • Mutually Exclusive (Disjoint): Events CANNOT happen at the same time ().
  • Independent: One event happening does NOT change the probability of the other.
    • Test: or .
  • Crucial Fact: If , then they cannot be both independent and mutually exclusive.

7.1 Conditional Probability: The “Reduced Sample Space”

is the probability of restricted to the universe of .

  • Formula: .
  • Multiplication Rule: .

8.1 Total Probability & Partitions

If states cover the entire space and don’t overlap (A Partition), the probability of any event is:

  • Visual: Think of as different factories and as a “defective item.”

8.2 Bayes Theorem (Belief Revision)

Bayes allows us to “invert” conditional probability.

  • Formula: .
  • Interpretation:
    • : Prior (What we knew before).
    • : Likelihood (How often the evidence happens in this state).
    • : Posterior (What we know after the evidence).

TIP

The Table Method for Bayes:

StatePriorLikelihoodProductPosterior (Product / Sum)
Sum1.0-1.0

📈 Volume IV: Random Variables & Statistical Blueprints

🧬 Week 9: Discrete Random Variables & PMFs

A Random Variable () is a function that maps outcomes of a random experiment to real numbers. It translates “Heads/Tails” into “1/0”.

9.1 The Probability Mass Function (PMF) -

For a discrete RV, the PMF gives the probability of taking a specific value .

  • Notation: .
  • Existence Constraints:
    1. for all .
    2. .

9.2 The Cumulative Distribution Function (CDF) -

The “Running Total” of probability.

  • Notation: .
  • Critical Properties:
    • is non-decreasing.
    • and .
    • .

9.3 Support of a Random Variable

The set of all values for which .

  • Finite Support: .
  • Countably Infinite Support: .

⚖️ Week 10: The Algebra of Expectation & Variance

To summarize a distribution, we use its Expected Value (The “Center”) and Variance (The “Spread”).

10.1 Expected Value ()

  • Definition: The long-term weighted average of outcomes.
  • Formula: .
  • Law of the Unconscious Statistician (LOTUS): To find the expectation of a function :
    • .

10.2 Linearity of Expectation (The “God” Rule)

Expectation is a linear operator. This holds regardless of whether variables are independent!

  • .
  • .

10.3 Variance ()

  • Definition: The expected value of the squared deviation from the mean.
  • Master Identity: .
  • Transformation: .
  • Std Dev: .

10.4 Independent Variables

If and are Independent:

  • Expectation: .
  • Variance: . (Note: It is always addition because squaring terms removes negative signs).
  • Covariance: .

🏗️ Week 11: Statistical Blueprints (Specific Discrete Distributions)

Certain experiment types appear so frequently that their distributions are pre-calculated.

11.1 The Bernoulli Trial ()

A single trial with two outcomes: Success (1) or Failure (0).

  • PMF: .
  • Stats: , .

11.2 The Binomial Distribution ()

The count of successes in Independent Bernoulli trials.

  • PMF: .
  • Conditions:
    1. Fixed number of trials .
    2. Each trial is independent.
    3. is constant ().
  • Stats: , .

11.3 The Poisson Distribution ()

Used for counts of “rare” events happening in a fixed interval (Time, Space, Volume).

  • PMF: .
  • Rate Scaling: If is for time , the rate for time is .
  • Stats: , .

IMPORTANT

The Binomial Poisson Approximation: When and , the Binomial distribution can be modeled as .


🌊 Week 12: Continuous Blueprints (Uniform & Exponential)

In the continuous domain, we deal with “densities” rather than discrete chunks of probability.

12.1 The Probability Density Function (PDF) -

For a continuous RV, is always 0. We only measure probability over an interval .

  • Core Relation: .
  • Constraint: Area under the entire curve .

12.2 The Continuous Uniform Distribution ()

Equally likely outcomes over a range .

  • PDF: for .
  • Stats: (Center), .

12.3 The Exponential Distribution ()

Models the time between events in a Poisson process.

  • PDF: for .
  • CDF: .
  • Stats: , .

TIP

The Memoryless Property: . “The probability that the car survives another hours given it has already survived hours is the same as the probability a brand new car survives hours.” Only the Exponential distribution has this property.

12.4 The Poisson-Exponential Duality

  • Poisson: Counts of events ( events in time ).
  • Exponential: Time until the next event ( time between events).
  • If events occur at rate (Poisson), the time between them follows .

🏁 Final Conclusion

This guide covers the entire spectrum of Statistics 1. From the raw classification of data to the sophisticated modeling of continuous processes, you now possess the complete theoretical toolkit for mastery.