MTH-361A | Spring 2025 | University of Portland
February 17, 2025
The Law of Large Numbers
It states that as the number of trials in a random experiment increases, the sample mean approaches the expected value.
Example Bernoulli Trials Simulation
Let \(p=0.60\) be the “success” probability of a Bernoulli r.v. \(X\), where \(\text{E}(X) = p\).
Exponential Distribution
Exponential R.V.
Let \(\lambda=\frac{1}{15}\) be the success probability.
\[ \begin{aligned} \text{R.V. } & \longrightarrow X \sim \text{Exp}\left(\frac{1}{15}\right) \\ \text{PDF } & \longrightarrow f(x) = \frac{1}{15} e^{- \frac{1}{15} x} \\ \text{for } & x \in [0,\infty) \end{aligned} \]
Exponential Distribution
Example:
What is the probability that “success” happens on 15 unit length or less, given \(\lambda=\frac{1}{15}\)? \[ \begin{aligned} P(X \le 15) & = \int_0^{15} f(x) \ dx \\ & = \int_0^{15} \frac{1}{15} e^{-\frac{1}{15} x} \ dx \\ P(X \le 15) & \approx 0.632 \end{aligned} \]
Using R:
## [1] 0.6321206
\(\star\) Note that the
pexp()
function computes the probability \(P(X \le x)\), meaning it computes the sum
of all probabilities from \(X=0\) to
\(X=x\) using the Exponential PDF. The
dexp()
function computes the density, not probability
because \(P(X = x)=0\) at any \(x\).
Exponential Distribution with Expected Value
Exponential R.V.
Let \(\lambda=\frac{1}{15}\) be the success probability.
\[ \begin{aligned} \text{R.V. } & \longrightarrow X \sim \text{Exp}\left(\frac{1}{15}\right) \\ \text{PDF } & \longrightarrow f(x) = \frac{1}{15} e^{- \frac{1}{15} x} \\ \text{for } & x \in [0,\infty) \\ \text{expected value} & \longrightarrow \text{E}(X) = 15 \end{aligned} \]
In general, the expected value of the exponential r.v. is given by \[\text{E}(X) = \frac{1}{\lambda},\] which is the reciprocal of the “success” rate.
Random Sampling from the Exponential Distribution
Sample Mean vs the Expected Value
The sample mean of \(0.88\) is not exactly equal to the expected value of \(1\) due to sampling variability. As we increase the number of samples, the sample mean gets closer to the expectation.
Geometric Random Sampling using R
Binomial Distribution
Binomial R.V.
Let \(p=0.50\) be the success probability and \(n=10\) the number of trials.
\[ \begin{aligned} \text{R.V. } & \longrightarrow X \sim \text{Binom}(p) \\ \text{PMF } & \longrightarrow P(X=k) = \binom{n}{k} p^k (1-p)^{n-k} \\ \text{for } & k = 0,1,2,3, \cdots, n \end{aligned} \]
Suppose we conduct an experiment of flipping \(n\) fair coins in a sequence, where \(n\) is an integer. The sample space \(S\) contains all possible sequences of \(H\) and \(T\). Number of possible outcomes is \(|S| = 2^n\).
Visualizing the possible outcomes using Pascal’s triangle
\(\star\) Key Idea: Pascal’s Triangle helps us visualize the total possible sequences of “success” (\(H\)) outcomes given \(n\) independent trials.
Let \(X\) be the r.v. that counts the number of \(H\) outcomes in \(n\) trials.
Pascal’s triangle helps us count
\(\dagger\) Can you determine the ways \(H\) can occur in \(4\) trials using Pascal’s triangle?
Compute the probability of observing a certain number of “success” (\(H\)) outcomes in \(n\) trials.
\(\dagger\) Can you determine the probabilities of observing \(H\) outcomes in \(4\) trials?
How many \(H\) outcomes do we expect to have in \(n\) independent Bernoulli trials?
Example
In general:
\(\star\) Key Idea: expected number of successes in \(n\) trials with success probability \(p\). Over many repetitions, the long-run average number of successes is \(n \times p\), reflecting the frequentist interpretation of probability.
The Normal Approximation
The binomial distribution can be approximated by a normal distribution with mean \(\mu\) and variance \(\sigma^2\) when:
Mean and Standard Deviation of the Normal Approximation
\(\star\) Key Idea: The Binomial distribution is approximately the normal distribution given large enough samples because of the Law of Large Numbers.
.pdf
file.