Study Notes

Probability and Statistics Exam - Cheatsheet

Mihail Barbosu
0 imports

Free · 2 imports included

Study Notes Preview

3 sections locked
Section 1

Probability and Statistics Exam - Cheatsheet

STUDY GUIDE

🎓 Probability and Statistics Exam - Study Guide

📋 Course Structure

code
📚 Probability and Statistics ├── 📖 Chapter 1: Expected Values │ ├── 🔹 Expected Value of a Function of Jointly Distributed Random Variables │ └── 🔹 Example: Expected Number of Seats Separating Individuals ├── 📖 Chapter 2: Covariance │ ├── 🔹 Definition and Interpretation of Covariance │ └── 🔹 Shortcut Formula for Covariance ├── 📖 Chapter 3: Correlation │ ├── 🔹 Definition of the Correlation Coefficient │ └── 🔹 Properties and Limitations of the Correlation Coefficient └── 📖 Chapter 4: The Bivariate Normal Distribution └── 🔹 Definition of the Bivariate Normal Distribution
Section 2

📖 Chapter 1: Expected Values

What this chapter covers: This chapter introduces the concept of expected values for functions of jointly distributed random variables. It differentiates between discrete and continuous random variables, providing the appropriate formulas for calculating expected values in each case. This lays the foundation for understanding more complex statistical measures.

🔑 Essential Concepts & Formulas

Concept/FormulaDefinition/EquationWhen to UseQuick Check
Expected Value (Discrete)E[h(X, Y)] = ΣΣ h(x, y) · p(x, y)X, Y are discreteSum of probabilities = 1
Expected Value (Continuous)E[h(X, Y)] = ∫∫ h(x, y) · f(x, y) dx dyX, Y are continuousIntegral of pdf = 1
h(x,y) in Seat Exampleh(x, y) =x-y- 1

🛠️ Problem Types

Type A: Calculating E[h(X,Y)] for Discrete Variables Setup: "When you see a joint probability mass function p(x,y) and a function h(x,y) for discrete variables." Method: Calculate E[h(X, Y)] = ΣΣ h(x, y) · p(x, y) over all possible values of x and y. Example: p(1,1) = 0.2, p(1,2) = 0.3, h(x,y) = x + y. E[h(X,Y)] = (1+1)*0.2 + (1+2)*0.3 = 0.4 + 0.9 = 1.3

Type B: Calculating E[h(X,Y)] for Continuous Variables Setup: "When you see a joint probability density function f(x,y) and a function h(x,y) for continuous variables." Method: Calculate E[h(X, Y)] = ∫∫ h(x, y) · f(x, y) dx dy over the entire range of x and y. Example: f(x,y) = 2xy for 0<x<1, 0<y<1, h(x,y) = xy. E[h(X,Y)] = ∫∫ (xy)*(2xy) dx dy = ∫∫ 2x²y² dx dy = 2 * (1/3) * (1/3) = 2/9

🧮 Solved Example

Problem: Five friends have tickets for seats 1-5. What is the expected number of seats separating any two of them?

Given: h(x, y) = |x-y| - 1, p(x, y) = 1/20 if x ≠ y, and 0 otherwise.

"
Solution: E[h(X, Y)] = ΣΣ h(x, y) · p(x, y) = (1/20) * ΣΣ |x-y| - 1. Calculate the sum over all x, y from 1 to 5 where x ≠ y. The sum of |x-y|-1 is 20. Therefore, E[h(X, Y)] = (1/20) * 20 = 1.
"
Answer: 1

⚠️ Common Mistakes

❌ Mistake 1: Using the wrong formula for discrete vs. continuous variables. ✅ How to avoid: Identify whether the variables are discrete or continuous before applying the formula.

❌ Mistake 2: Incorrectly calculating the joint probability mass function or density function. ✅ How to avoid: Ensure the probabilities sum to 1 for discrete variables and the integral equals 1 for continuous variables.

🦁 Erik's Tip

When calculating expected values, always double-check that your function h(x,y) accurately represents the quantity you're trying to find the expected value of. A small error in h(x,y) can lead to a completely wrong answer.

📖 Chapter 2: Covariance

What this chapter covers: This chapter defines covariance as a measure of the relationship between two random variables. It provides both the conceptual definition and a shortcut formula for calculating covariance. The chapter also discusses the interpretation of covariance values.

🔑 Essential Concepts & Formulas

Concept/FormulaDefinition/EquationWhen to UseQuick Check
Covariance (Definition)Cov(X, Y) = E[(X - μX)(Y - μY)]Conceptual understandingCheck if positive/negative makes sense
Covariance (Shortcut)Cov(X, Y) = E(XY) - μX · μYEasier calculationCompare with definition result
μXE[X]Calculating mean of XSum/Integral of xp(x) or xf(x)
μYE[Y]Calculating mean of YSum/Integral of yp(y) or yf(y)

🛠️ Problem Types

Type A: Calculating Covariance using the Definition Setup: "When you have the joint distribution and the means of X and Y." Method: Calculate Cov(X, Y) = E[(X - μX)(Y - μY)] = ΣΣ (x - μX)(y - μY) * p(x,y) for discrete variables or ∫∫ (x - μX)(y - μY) * f(x,y) dx dy for continuous variables. Example: μX = 2, μY = 3, p(2,3) = 0.5, p(3,4) = 0.5. Cov(X,Y) = (2-2)(3-3)*0.5 + (3-2)(4-3)*0.5 = 0 + 0.5 = 0.5

Type B: Calculating Covariance using the Shortcut Formula Setup: "When you have E(XY), μX, and μY." Method: Calculate Cov(X, Y) = E(XY) - μX · μY. Example: E(XY) = 7, μX = 2, μY = 3. Cov(X, Y) = 7 - 2*3 = 1

🧮 Solved Example

Problem: Calculate the covariance of X and Y given E(XY) = 10, μX = 2, and μY = 4.

Given: E(XY) = 10, μX = 2, μY = 4

"
Solution: Cov(X, Y) = E(XY) - μX · μY = 10 - 2 * 4 = 10 - 8 = 2
"
Answer: 2

⚠️ Common Mistakes

❌ Mistake 1: Forgetting to subtract the means when using the definition of covariance. ✅ How to avoid: Double-check that you are calculating E[(X - μX)(Y - μY)] and not just E[XY].

❌ Mistake 2: Calculating E(XY) incorrectly. ✅ How to avoid: Ensure you are using the correct joint distribution and summing/integrating over all possible values.

🦁 Erik's Tip

The shortcut formula for covariance is almost always easier to use than the definition, especially when dealing with complex joint distributions. Make sure you can quickly calculate E(XY), μX, and μY.

📖 Chapter 3: Correlation

What this chapter covers: This chapter introduces the correlation coefficient as a normalized measure of the linear relationship between two random variables. It defines the correlation coefficient and discusses its properties, including its range and interpretation.

🔑 Essential Concepts & Formulas

Concept/FormulaDefinition/EquationWhen to UseQuick Check
Correlation CoefficientρX,Y = Cov(X, Y) / (σX · σY)Measuring linear relationship-1 ≤ ρ ≤ 1
Standard Deviation (σX)√Var(X)Calculating standard deviation of XσX ≥ 0
Standard Deviation (σY)√Var(Y)Calculating standard deviation of YσY ≥ 0
Variance (Var(X))E[(X - μX)²] or E[X²] - (E[X])²Calculating variance of XVar(X) ≥ 0

🛠️ Problem Types

Type A: Calculating the Correlation Coefficient Setup: "When you have Cov(X, Y), σX, and σY." Method: Calculate ρX,Y = Cov(X, Y) / (σX · σY). Example: Cov(X, Y) = 5, σX = 2, σY = 3. ρX,Y = 5 / (2 * 3) = 5/6 ≈ 0.833

Type B: Interpreting the Correlation Coefficient Setup: "When you are given a value of ρX,Y." Method: If ρX,Y is close to 1, there is a strong positive linear relationship. If ρX,Y is close to -1, there is a strong negative linear relationship. If ρX,Y is close to 0, there is little or no linear relationship. Example: ρX,Y = 0.9 indicates a strong positive linear relationship.

🧮 Solved Example

Problem: Calculate the correlation coefficient given Cov(X, Y) = 8, σX = 4, and σY = 2.

Given: Cov(X, Y) = 8, σX = 4, σY = 2

"
Solution: ρX,Y = Cov(X, Y) / (σX · σY) = 8 / (4 * 2) = 8 / 8 = 1
"
Answer: 1

⚠️ Common Mistakes

❌ Mistake 1: Forgetting to divide by the standard deviations when calculating the correlation coefficient. ✅ How to avoid: Double-check that you are calculating ρX,Y = Cov(X, Y) / (σX · σY) and not just using the covariance.

❌ Mistake 2: Assuming that a correlation of 0 implies independence. ✅ How to avoid: Remember that a correlation of 0 only implies no linear relationship. There may be a non-linear relationship.

🦁 Erik's Tip

Always remember that correlation does not imply causation! Just because two variables are highly correlated doesn't mean that one causes the other. There may be a lurking variable or the relationship may be coincidental.

3 more sections

Create a free account to import and read the full study notes — all 5 sections.

No credit card · 2 free imports included

    Probability and Statistics Exam - Cheatsheet — Cheatsheet | Evrika | Evrika Study