@@ -111,19 +111,32 @@ To appreciate how statisticians connect probabilities to data, the key is to und
111111
112112**Scalar example**
113113
114+ Let $X$ be a scalar random variable that takes on the $I$ possible values
115+ $0, 1, 2, \ldots, I-1$ with probabilities
114116
115- Consider the following discrete distribution
117+ $$
118+ {\rm Prob}(X = i) = f_i, \quad
119+ $$
120+ where
121+
122+ $$
123+ f_i \geqslant 0, \quad \sum_i f_i = 1 .
124+ $$
125+
126+ We sometimes write
116127
117128$$
118- X \sim \{ {f_i}\} _ {i=0}^{I-1},\quad f_i \geqslant 0, \quad \sum_i f_i = 1
129+ X \sim \{ {f_i}\} _ {i=0}^{I-1}
119130$$
120131
121- Draw a sample $x_0, x_1, \dots , x_{N-1}$, $N$ draws of $X$ from $\{f_i\}^I_{i=1}$.
132+ as a short-hand way of saying that the random variable $X$ is described by the probability distribution $ \{{f_i}\}_{i=0}^{I-1}$.
133+
134+ Consider drawing a sample $x_0, x_1, \dots , x_{N-1}$ of $N$ independent and identically distributoed draws of $X$.
122135
123136What do the "identical" and "independent" mean in IID or iid ("identically and independently distributed)?
124137
125138- "identical" means that each draw is from the same distribution.
126- - "independent" means that the joint distribution equal tthe product of marginal distributions, i.e.,
139+ - "independent" means that joint distribution equal products of marginal distributions, i.e.,
127140
128141$$
129142\begin{aligned}
132145\end{aligned}
133146$$
134147
135- Consider the **empirical distribution**:
148+ We define an e **empirical distribution** as follows.
149+
150+ For each $i = 0,\dots,I-1$, let
136151
137152$$
138153\begin{aligned}
139- i & = 0,\dots,I-1,\\
140154N_i & = \text{number of times} \ X = i,\\
141155N & = \sum^{I-1}_ {i=0} N_i \quad \text{total number of draws},\\
142156\tilde {f_i} & = \frac{N_i}{N} \sim \ \text{frequency of draws for which}\ X=i
@@ -425,7 +439,7 @@ Conditional distributions are
425439
426440$$
427441\begin{aligned}
428- \textrm{Prob}\{ X=i|Y=j\} & =\frac{f_ig_j}{\sum_ {i}f_ig_j}=\frac{f_ig_j}{g_i }=f_i \\
442+ \textrm{Prob}\{ X=i|Y=j\} & =\frac{f_ig_j}{\sum_ {i}f_ig_j}=\frac{f_ig_j}{g_j }=f_i \\
429443\textrm{Prob}\{ Y=j|X=i\} & =\frac{f_ig_j}{\sum_ {j}f_ig_j}=\frac{f_ig_j}{f_i}=g_j
430444\end{aligned}
431445$$
609623\begin{aligned}
610624\tilde{U} & =F(X)=1-\lambda^{x+1}\\
6116251-\tilde{U} & =\lambda^{x+1}\\
612- log(1-\tilde{U})& =(x+1)\log\lambda\\
626+ \ log(1-\tilde{U})& =(x+1)\log\lambda\\
613627\frac{\log(1-\tilde{U})}{\log\lambda}& =x+1\\
614628\frac{\log(1-\tilde{U})}{\log\lambda}-1 &=x
615629\end{aligned}
@@ -1561,7 +1575,7 @@ Now we'll try to go in a reverse direction.
15611575
15621576We'll find that from two marginal distributions, can we usually construct more than one joint distribution that verifies these marginals.
15631577
1564- Each of these joint distributions is called a ** coupling** of the two martingal distributions.
1578+ Each of these joint distributions is called a ** coupling** of the two marginal distributions.
15651579
15661580Let's start with marginal distributions
15671581
0 commit comments