Die Linearität der Erwartung hält wenn \(X_1,\ldots,X_n\) nicht unabhängig, du dummbatzi sind?
Note 1: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
added
Note Type: Horvath Cloze
GUID:
LE!#OdHaHH
Previous
Note did not exist
New Note
Front
Back
Die Linearität der Erwartung hält wenn \(X_1,\ldots,X_n\) nicht unabhängig, du dummbatzi sind?
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | Die Linearität der Erwartung hält wenn \(X_1,\ldots,X_n\) {{c1::nicht unabhängig, du dummbatzi}} sind? |
Note 2: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_06f8d72e
Before
Front
The density function (Dichtefunktion) of a random variable \(X\) is:
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
Back
The density function (Dichtefunktion) of a random variable \(X\) is:
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
It is zero outside \(W_X\). The density function the random variable's distribution.
After
Front
The density function (Dichtefunktion) of a random variable \(X\) is:
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
Back
The density function (Dichtefunktion) of a random variable \(X\) is:
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
\[ {{c1:: f_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X=x]}}. \]
It is zero outside \(W_X\). The density function uniquely determines the random variable's distribution.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Extra | It is zero outside \(W_X\). The density function |
It is zero outside \(W_X\). The density function uniquely determines the random variable's distribution. |
Note 3: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Classic
GUID:
modified
Note Type: Horvath Classic
GUID:
sc_0aab872f
Before
Front
Warum muss in Def. 2.22 die Produktregel für alle Teilmengen gelten, und nicht nur für die paarweisen oder den vollen Schnitt?
Back
Warum muss in Def. 2.22 die Produktregel für alle Teilmengen gelten, und nicht nur für die paarweisen oder den vollen Schnitt?
Paarweise \(\not \implies\) vollen Schnitt: Zwei faire Münzen, \(A=\)„\(M_1\) Kopf", \(B=\)„\(M_2\) Kopf", \(C=\)„Ergebnisse verschieden": je zwei Ereignisse sind unabhängig, aber \(\Pr[A\cap B\cap C]=0\neq\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\).
Voller Schnitt \(\not \implies\) paarweise: Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).
Beide Bedingungen zusammen sind nötig — daher fordert Def. 2.22 die Produktregel für alle nichtleeren Teilmengen.
Voller Schnitt \(\not \implies\) paarweise: Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).
Beide Bedingungen zusammen sind nötig — daher fordert Def. 2.22 die Produktregel für alle nichtleeren Teilmengen.
After
Front
Warum muss für Unabhängigkeit die Produktregel für alle Teilmengen gelten, und nicht nur für die paarweisen oder den vollen Schnitt?
Back
Warum muss für Unabhängigkeit die Produktregel für alle Teilmengen gelten, und nicht nur für die paarweisen oder den vollen Schnitt?
Paarweise \(\not \implies\) vollen Schnitt: Zwei faire Münzen, \(A=\)„\(M_1\) Kopf", \(B=\)„\(M_2\) Kopf", \(C=\)„Ergebnisse verschieden": je zwei Ereignisse sind unabhängig, aber \(\Pr[A\cap B\cap C]=0\neq\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\).
Voller Schnitt \(\not \implies\) paarweise: Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).
Beide Bedingungen zusammen sind nötig — daher fordert Unabhängigkeit die Produktregel für alle nichtleeren Teilmengen.
Voller Schnitt \(\not \implies\) paarweise: Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).
Beide Bedingungen zusammen sind nötig — daher fordert Unabhängigkeit die Produktregel für alle nichtleeren Teilmengen.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Front | Warum muss |
Warum muss für Unabhängigkeit die Produktregel für <strong>alle Teilmengen</strong> gelten, und nicht nur für die paarweisen oder den vollen Schnitt? |
| Back | <strong>Paarweise \(\not \implies\) vollen Schnitt:</strong> Zwei faire Münzen, \(A=\)„\(M_1\) Kopf", \(B=\)„\(M_2\) Kopf", \(C=\)„Ergebnisse verschieden": je zwei Ereignisse sind unabhängig, aber \(\Pr[A\cap B\cap C]=0\neq\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\).<br><strong>Voller Schnitt \(\not \implies\) paarweise:</strong> Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).<br><br>Beide Bedingungen zusammen sind nötig — daher fordert |
<strong>Paarweise \(\not \implies\) vollen Schnitt:</strong> Zwei faire Münzen, \(A=\)„\(M_1\) Kopf", \(B=\)„\(M_2\) Kopf", \(C=\)„Ergebnisse verschieden": je zwei Ereignisse sind unabhängig, aber \(\Pr[A\cap B\cap C]=0\neq\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\).<br><br><strong>Voller Schnitt \(\not \implies\) paarweise:</strong> Zufallszahl in \(\{1,\ldots,8\}\), \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\): \(\Pr[A\cap B\cap C]=\tfrac{1}{8}=\Pr[A]\Pr[B]\Pr[C]\), aber \(\Pr[A\cap B]=\tfrac{1}{8}\neq\tfrac{1}{4}=\Pr[A]\Pr[B]\).<br><br>Beide Bedingungen zusammen sind nötig — daher fordert Unabhängigkeit die Produktregel für <strong>alle</strong> nichtleeren Teilmengen. |
Note 4: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_1de913c2
Before
Front
(Satz 2.33 — Linearität) For \(X=a_1X_1+\cdots+a_nX_n+b\):
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}. \]
Holds even if \(X_1,\ldots,X_n\) are not independent. Proof Included
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}. \]
Holds even if \(X_1,\ldots,X_n\) are not independent. Proof Included
Back
(Satz 2.33 — Linearität) For \(X=a_1X_1+\cdots+a_nX_n+b\):
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}. \]
Holds even if \(X_1,\ldots,X_n\) are not independent. Proof Included
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}. \]
Holds even if \(X_1,\ldots,X_n\) are not independent. Proof Included
Proof:
Using Lemma 2.29 (\(\mathbb{E}[X]=\sum_\omega X(\omega)\Pr[\omega]\)):
\[ \mathbb{E}[X]=\sum_\omega(a_1X_1(\omega)+\cdots+a_nX_n(\omega)+b)\Pr[\omega] =a_1\underbrace{\sum_\omega X_1(\omega)\Pr[\omega]}_{=\mathbb{E}[X_1]}+\cdots+b\underbrace{\sum_\omega\Pr[\omega]}_{=1}.\quad\square \]
The key is that the outer sum \(\sum_\omega\) distributes over the linear combination.
Using Lemma 2.29 (\(\mathbb{E}[X]=\sum_\omega X(\omega)\Pr[\omega]\)):
\[ \mathbb{E}[X]=\sum_\omega(a_1X_1(\omega)+\cdots+a_nX_n(\omega)+b)\Pr[\omega] =a_1\underbrace{\sum_\omega X_1(\omega)\Pr[\omega]}_{=\mathbb{E}[X_1]}+\cdots+b\underbrace{\sum_\omega\Pr[\omega]}_{=1}.\quad\square \]
The key is that the outer sum \(\sum_\omega\) distributes over the linear combination.
After
Front
(Linearität) For \(X=a_1X_1+\cdots+a_nX_n+b\):
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}\]
Proof Included
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}\]
Proof Included
Back
(Linearität) For \(X=a_1X_1+\cdots+a_nX_n+b\):
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}\]
Proof Included
\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}\]
Proof Included
Holds even if \(X_1,\ldots,X_n\) are not independent.
Proof:
Using Lemma 2.29 (\(\mathbb{E}[X]=\sum_\omega X(\omega)\Pr[\omega]\)):
\[ \mathbb{E}[X]=\sum_\omega(a_1X_1(\omega)+\cdots+a_nX_n(\omega)+b)\Pr[\omega] =a_1\underbrace{\sum_\omega X_1(\omega)\Pr[\omega]}_{=\mathbb{E}[X_1]}+\cdots+b\underbrace{\sum_\omega\Pr[\omega]}_{=1}.\quad\square \]
The key is that the outer sum \(\sum_\omega\) distributes over the linear combination.
Proof:
Using Lemma 2.29 (\(\mathbb{E}[X]=\sum_\omega X(\omega)\Pr[\omega]\)):
\[ \mathbb{E}[X]=\sum_\omega(a_1X_1(\omega)+\cdots+a_nX_n(\omega)+b)\Pr[\omega] =a_1\underbrace{\sum_\omega X_1(\omega)\Pr[\omega]}_{=\mathbb{E}[X_1]}+\cdots+b\underbrace{\sum_\omega\Pr[\omega]}_{=1}.\quad\square \]
The key is that the outer sum \(\sum_\omega\) distributes over the linear combination.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | ( |
(<b>Linearität</b>) For \(X=a_1X_1+\cdots+a_nX_n+b\):<br>\[ \mathbb{E}[X] = {{c1::a_1\mathbb{E}[X_1]+\cdots+a_n\mathbb{E}[X_n]+b}}\]<br> <em>Proof Included</em> |
| Extra | Holds even if \(X_1,\ldots,X_n\) are <strong>not independent</strong>.<strong><br><br>Proof:</strong><br>Using Lemma 2.29 (\(\mathbb{E}[X]=\sum_\omega X(\omega)\Pr[\omega]\)):<br>\[ \mathbb{E}[X]=\sum_\omega(a_1X_1(\omega)+\cdots+a_nX_n(\omega)+b)\Pr[\omega] =a_1\underbrace{\sum_\omega X_1(\omega)\Pr[\omega]}_{=\mathbb{E}[X_1]}+\cdots+b\underbrace{\sum_\omega\Pr[\omega]}_{=1}.\quad\square \]<br>The key is that the outer sum \(\sum_\omega\) distributes over the linear combination. |
Note 5: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_1eb62870
Before
Front
(Satz 2.32 — Gesetz der totalen Erwartung) Let \(A_1,\ldots,A_n\) partition \(\Omega\) with all \(\Pr[A_i]>0\). Then:
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Back
(Satz 2.32 — Gesetz der totalen Erwartung) Let \(A_1,\ldots,A_n\) partition \(\Omega\) with all \(\Pr[A_i]>0\). Then:
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Proof:
\[ \mathbb{E}[X]=\sum_{x}x\cdot\Pr[X=x]\overset{\text{total prob.}}{=}\sum_x x\sum_i\Pr[X=x|A_i]\Pr[A_i] =\sum_i\Pr[A_i]\underbrace{\sum_x x\Pr[X=x|A_i]}_{=\mathbb{E}[X|A_i]}.\quad\square \]
(Uses the law of total probability to expand \(\Pr[X=x]\), then swaps summation order.)
\[ \mathbb{E}[X]=\sum_{x}x\cdot\Pr[X=x]\overset{\text{total prob.}}{=}\sum_x x\sum_i\Pr[X=x|A_i]\Pr[A_i] =\sum_i\Pr[A_i]\underbrace{\sum_x x\Pr[X=x|A_i]}_{=\mathbb{E}[X|A_i]}.\quad\square \]
(Uses the law of total probability to expand \(\Pr[X=x]\), then swaps summation order.)
After
Front
(Gesetz der totalen Erwartung, not script) Let \(A_1,\ldots,A_n\) partition \(\Omega\) with all \(\Pr[A_i]>0\). Then:
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Back
(Gesetz der totalen Erwartung, not script) Let \(A_1,\ldots,A_n\) partition \(\Omega\) with all \(\Pr[A_i]>0\). Then:
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Proof:
\[\begin{align} \mathbb{E}[X] &=\sum_{x}x\cdot\Pr[X=x] \\ &\overset{\text{total prob}}{=}\sum_x x\sum_i\Pr[X=x|A_i]\Pr[A_i] \\ &=\sum_i\Pr[A_i]\underbrace{\sum_x x\Pr[X=x|A_i]}_{=\mathbb{E}[X|A_i]} \end{align}\]
(Uses the law of total probability to expand \(\Pr[X=x]\), then swaps summation order.)
\[\begin{align} \mathbb{E}[X] &=\sum_{x}x\cdot\Pr[X=x] \\ &\overset{\text{total prob}}{=}\sum_x x\sum_i\Pr[X=x|A_i]\Pr[A_i] \\ &=\sum_i\Pr[A_i]\underbrace{\sum_x x\Pr[X=x|A_i]}_{=\mathbb{E}[X|A_i]} \end{align}\]
(Uses the law of total probability to expand \(\Pr[X=x]\), then swaps summation order.)
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | <b>(Gesetz der totalen Erwartung, not script)</b> Let \(A_1,\ldots,A_n\) partition \(\Omega\) with all \(\Pr[A_i]>0\). Then:<br>\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{n}\mathbb{E}[X|A_i]\cdot\Pr[A_i]}}. \]<br><em>Proof Included</em> | |
| Extra | <strong>Proof:</strong><br>\[ \mathbb{E}[X]=\sum_{x}x\cdot\Pr[X=x]\overset{\text{total prob |
<strong>Proof:</strong><br>\[\begin{align} \mathbb{E}[X] &=\sum_{x}x\cdot\Pr[X=x] \\ &\overset{\text{total prob}}{=}\sum_x x\sum_i\Pr[X=x|A_i]\Pr[A_i] \\ &=\sum_i\Pr[A_i]\underbrace{\sum_x x\Pr[X=x|A_i]}_{=\mathbb{E}[X|A_i]} \end{align}\]<br>(Uses the law of total probability to expand \(\Pr[X=x]\), then swaps summation order.) |
Note 6: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_3121875f
Before
Front
(Definition 2.22) Events \(A_1,\ldots,A_n\) are called independent (unabhängig) if for all subsets
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\}\):
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}. \tag{2.2} \]
An infinite family \((A_i)_{i\in\mathbb{N}}\) is independent if (2.2) holds for {{c2::every finite
subset \(I\subseteq\mathbb{N}\)}}.
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\}\):
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}. \tag{2.2} \]
An infinite family \((A_i)_{i\in\mathbb{N}}\) is independent if (2.2) holds for {{c2::every finite
subset \(I\subseteq\mathbb{N}\)}}.
Back
(Definition 2.22) Events \(A_1,\ldots,A_n\) are called independent (unabhängig) if for all subsets
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\}\):
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}. \tag{2.2} \]
An infinite family \((A_i)_{i\in\mathbb{N}}\) is independent if (2.2) holds for {{c2::every finite
subset \(I\subseteq\mathbb{N}\)}}.
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\}\):
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}. \tag{2.2} \]
An infinite family \((A_i)_{i\in\mathbb{N}}\) is independent if (2.2) holds for {{c2::every finite
subset \(I\subseteq\mathbb{N}\)}}.
The definition requires the product rule for every non-empty sub-intersection — not just pairwise intersections, and not just the full \(n\)-fold intersection. Both conditions alone are insufficient:
Pairwise \(\not \implies\) full intersection: two fair coins, \(A=\)"\(M_1\) heads", \(B=\)"\(M_2\) heads", \(C=\)"results differ" are pairwise independent but \(\Pr[A\cap B\cap C]=0\neq\Pr[A]\Pr[B]\Pr[C]\).
Full intersection \(\not \implies\) pairwise: \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\) in \(\{1,\ldots,8\}\): \(\Pr[A\cap B\cap C]=\Pr[A]\Pr[B]\Pr[C]\) but \(A,B\) are not independent.
The total number of conditions to check for \(n\) events is \(2^n-1\) (all non-empty subsets of \(\{1,\ldots,n\}\)).
Pairwise \(\not \implies\) full intersection: two fair coins, \(A=\)"\(M_1\) heads", \(B=\)"\(M_2\) heads", \(C=\)"results differ" are pairwise independent but \(\Pr[A\cap B\cap C]=0\neq\Pr[A]\Pr[B]\Pr[C]\).
Full intersection \(\not \implies\) pairwise: \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\) in \(\{1,\ldots,8\}\): \(\Pr[A\cap B\cap C]=\Pr[A]\Pr[B]\Pr[C]\) but \(A,B\) are not independent.
The total number of conditions to check for \(n\) events is \(2^n-1\) (all non-empty subsets of \(\{1,\ldots,n\}\)).
After
Front
(Definition 2.22) Events \(A_1,\ldots,A_n\) are called independent (unabhängig) if {{c1::for all subsets
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\} \)}}:
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}\]
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\} \)}}:
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}\]
Back
(Definition 2.22) Events \(A_1,\ldots,A_n\) are called independent (unabhängig) if {{c1::for all subsets
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\} \)}}:
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}\]
\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\} \)}}:
\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}\]
The definition requires the product rule for every non-empty sub-intersection — not just pairwise intersections, and not just the full \(n\)-fold intersection. Both conditions alone are insufficient:
The total number of conditions to check for \(n\) events is \(2^n-1\) (all non-empty subsets of \(\{1,\ldots,n\}\)).
- Pairwise \(\not \implies\) full intersection: two fair coins, \(A=\)"\(M_1\) heads", \(B=\)"\(M_2\) heads", \(C=\)"results differ" are pairwise independent but \(\Pr[A\cap B\cap C]=0\neq\Pr[A]\Pr[B]\Pr[C]\).
- Full intersection \(\not \implies\) pairwise: \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\) in \(\{1,\ldots,8\}\): \(\Pr[A\cap B\cap C]=\Pr[A]\Pr[B]\Pr[C]\) but \(A,B\) are not independent.
The total number of conditions to check for \(n\) events is \(2^n-1\) (all non-empty subsets of \(\{1,\ldots,n\}\)).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | (Definition 2.22) Events \(A_1,\ldots,A_n\) are called <strong>independent</strong> (unabhängig) if for <strong>all</strong> subsets<br>\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\}\):<br>\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}} |
(Definition 2.22) Events \(A_1,\ldots,A_n\) are called <strong>independent</strong> (unabhängig) if {{c1::for <strong>all</strong> subsets<br>\(I\subseteq\{1,\ldots,n\}\) with \(I=\{i_1,\ldots,i_k\} \)}}:<br>\[ \Pr\!\left[A_{i_1}\cap\cdots\cap A_{i_k}\right] = {{c1::\Pr[A_{i_1}]\cdots\Pr[A_{i_k}]}}\] |
| Extra | The definition requires the product rule for <strong>every</strong> non-empty sub-intersection — not just pairwise intersections, and not just the full \(n\)-fold intersection. Both conditions alone are insufficient:<br>Pairwise \(\not \implies\) full intersection: two fair coins, \(A=\)"\(M_1\) heads", \(B=\)"\(M_2\) heads", \(C=\)"results differ" are pairwise independent but \(\Pr[A\cap B\cap C]=0\neq\Pr[A]\Pr[B]\Pr[C]\).< |
The definition requires the product rule for <strong>every</strong> non-empty sub-intersection — not just pairwise intersections, and not just the full \(n\)-fold intersection. Both conditions alone are insufficient:<br><br><ul><li>Pairwise \(\not \implies\) full intersection: two fair coins, \(A=\)"\(M_1\) heads", \(B=\)"\(M_2\) heads", \(C=\)"results differ" are pairwise independent but \(\Pr[A\cap B\cap C]=0\neq\Pr[A]\Pr[B]\Pr[C]\).</li><li>Full intersection \(\not \implies\) pairwise: \(A=\{1,2,3,4\}\), \(B=\{1,5,6,7\}\), \(C=B\) in \(\{1,\ldots,8\}\): \(\Pr[A\cap B\cap C]=\Pr[A]\Pr[B]\Pr[C]\) but \(A,B\) are not independent.</li></ul><br>The total number of conditions to check for \(n\) events is \(2^n-1\) (all non-empty subsets of \(\{1,\ldots,n\}\)).<br> |
Note 7: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_32a73428
Before
Front
(Satz von Bayes, Satz 2.15) Let \(A_1,\ldots,A_n\) be pairwise disjoint, \(B\subseteq\bigcup A_i\), \(\Pr[B]>0\). Then for any \(i\):
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
Back
(Satz von Bayes, Satz 2.15) Let \(A_1,\ldots,A_n\) be pairwise disjoint, \(B\subseteq\bigcup A_i\), \(\Pr[B]>0\). Then for any \(i\):
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
Proof: By definition \(\Pr[A_i|B]=\Pr[A_i\cap B]/\Pr[B]\). Numerator: \(\Pr[A_i\cap B]=\Pr[B|A_i]\cdot\Pr[A_i]\). Denominator: \(\Pr[B]=\sum_j\Pr[B|A_j]\Pr[A_j]\) (total probability). \(\square\)
Key use: "Invert" the direction of conditioning — from \(\Pr[B|A_i]\) (easy to measure) to \(\Pr[A_i|B]\) (what we want to know).
Key use: "Invert" the direction of conditioning — from \(\Pr[B|A_i]\) (easy to measure) to \(\Pr[A_i|B]\) (what we want to know).
After
Front
(Satz von Bayes) Let \(A_1,\ldots,A_n\) be pairwise disjoint, \(B\subseteq\bigcup A_i\), \(\Pr[B]>0\). Then for any \(i\):
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
Back
(Satz von Bayes) Let \(A_1,\ldots,A_n\) be pairwise disjoint, \(B\subseteq\bigcup A_i\), \(\Pr[B]>0\). Then for any \(i\):
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]
Proof Included
Proof: By definition \(\Pr[A_i|B]=\Pr[A_i\cap B]/\Pr[B]\). Numerator: \(\Pr[A_i\cap B]=\Pr[B|A_i]\cdot\Pr[A_i]\). Denominator: \(\Pr[B]=\sum_j\Pr[B|A_j]\Pr[A_j]\) (total probability). \(\square\)
Key use: "Invert" the direction of conditioning — from \(\Pr[B|A_i]\) (easy to measure) to \(\Pr[A_i|B]\) (what we want to know).
Key use: "Invert" the direction of conditioning — from \(\Pr[B|A_i]\) (easy to measure) to \(\Pr[A_i|B]\) (what we want to know).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | (Satz von Bayes |
(Satz von <b>Bayes</b>) Let \(A_1,\ldots,A_n\) be pairwise disjoint, \(B\subseteq\bigcup A_i\), \(\Pr[B]>0\). Then for any \(i\):<br>\[ \Pr[A_i|B] = {{c1::\frac{\Pr[B|A_i]\cdot\Pr[A_i]}{\sum_{j=1}^{n}\Pr[B|A_j]\cdot\Pr[A_j]} }}. \]<br><em>Proof Included</em> |
Note 8: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_35dcd411
Before
Front
(Expected Value as Sum) For any random variable \(X\):
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::sum definition}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::sum definition}}. \]
Proof Included
Back
(Expected Value as Sum) For any random variable \(X\):
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::sum definition}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::sum definition}}. \]
Proof Included
Proof:
\[ \mathbb{E}[X]=\sum_{x\in W_X}x\cdot\Pr[X=x]=\sum_{x\in W_X}x\cdot\sum_{\omega: X(\omega)=x}\Pr[\omega]=\sum_{\omega\in\Omega}X(\omega)\cdot\Pr[\omega].\quad\square \]
(Switch the order of summation: group by \(\omega\) instead of by value \(x\).)
\[ \mathbb{E}[X]=\sum_{x\in W_X}x\cdot\Pr[X=x]=\sum_{x\in W_X}x\cdot\sum_{\omega: X(\omega)=x}\Pr[\omega]=\sum_{\omega\in\Omega}X(\omega)\cdot\Pr[\omega].\quad\square \]
(Switch the order of summation: group by \(\omega\) instead of by value \(x\).)
After
Front
(Expected Value as Sum) For any random variable \(X\):
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::weighted sum definition}} \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::weighted sum definition}} \]
Proof Included
Back
(Expected Value as Sum) For any random variable \(X\):
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::weighted sum definition}} \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::weighted sum definition}} \]
Proof Included
Proof:
\[\begin{align} \mathbb{E}[X] &= \sum_{x\in W_X}x\cdot\Pr[X=x] \\ &=\sum_{x\in W_X}x\cdot\sum_{\omega: X(\omega)=x}\Pr[\omega] \\&=\sum_{\omega\in\Omega}X(\omega)\cdot\Pr[\omega].\quad \end{align}\]
(Switch the order of summation: group by \(\omega\) instead of by value \(x\).)
\[\begin{align} \mathbb{E}[X] &= \sum_{x\in W_X}x\cdot\Pr[X=x] \\ &=\sum_{x\in W_X}x\cdot\sum_{\omega: X(\omega)=x}\Pr[\omega] \\&=\sum_{\omega\in\Omega}X(\omega)\cdot\Pr[\omega].\quad \end{align}\]
(Switch the order of summation: group by \(\omega\) instead of by value \(x\).)
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | (Expected Value as Sum) For any random variable \(X\):<br>\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::sum definition}} |
(<b>Expected Value as Sum</b>) For any random variable \(X\):<br>\[ \mathbb{E}[X] = {{c1::\sum_{\omega\in\Omega} X(\omega)\cdot\Pr[\omega]::weighted sum definition}} \]<br><em>Proof Included</em> |
| Extra | <strong>Proof:</strong><br>\[ \mathbb{E}[X] |
<strong>Proof:</strong><br>\[\begin{align} \mathbb{E}[X] &= \sum_{x\in W_X}x\cdot\Pr[X=x] \\ &=\sum_{x\in W_X}x\cdot\sum_{\omega: X(\omega)=x}\Pr[\omega] \\&=\sum_{\omega\in\Omega}X(\omega)\cdot\Pr[\omega].\quad \end{align}\]<br>(Switch the order of summation: group by \(\omega\) instead of by value \(x\).) |
Note 9: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_38dbfe49
Before
Front
(Expected Value) For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i]}}. \]
Proof Included
Back
(Expected Value) For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i]}}. \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i]}}. \]
Proof Included
Proof:
\[ \mathbb{E}[X]=\sum_{i=0}^{\infty}i\cdot\Pr[X=i]=\sum_{i=0}^{\infty}\sum_{j=1}^{i}\Pr[X=i]=\sum_{j=1}^{\infty}\sum_{i=j}^{\infty}\Pr[X=i]=\sum_{j=1}^{\infty}\Pr[X\ge j].\quad\square \]
(The key step is swapping the order of summation: instead of summing over \(i\) and counting 1 for each \(j\le i\), sum over \(j\) and count all \(i\ge j\).)
\[ \mathbb{E}[X]=\sum_{i=0}^{\infty}i\cdot\Pr[X=i]=\sum_{i=0}^{\infty}\sum_{j=1}^{i}\Pr[X=i]=\sum_{j=1}^{\infty}\sum_{i=j}^{\infty}\Pr[X=i]=\sum_{j=1}^{\infty}\Pr[X\ge j].\quad\square \]
(The key step is swapping the order of summation: instead of summing over \(i\) and counting 1 for each \(j\le i\), sum over \(j\) and count all \(i\ge j\).)
After
Front
(Expected Value) For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i] :: bound form}} \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i] :: bound form}} \]
Proof Included
Back
(Expected Value) For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i] :: bound form}} \]
Proof Included
\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i] :: bound form}} \]
Proof Included
Proof:
\[ \mathbb{E}[X]=\sum_{i=0}^{\infty}i\cdot\Pr[X=i]=\sum_{i=0}^{\infty}\sum_{j=1}^{i}\Pr[X=i]=\sum_{j=1}^{\infty}\sum_{i=j}^{\infty}\Pr[X=i]=\sum_{j=1}^{\infty}\Pr[X\ge j].\quad\square \]
(The key step is swapping the order of summation: instead of summing over \(i\) and counting 1 for each \(j\le i\), sum over \(j\) and count all \(i\ge j\).)
\[ \mathbb{E}[X]=\sum_{i=0}^{\infty}i\cdot\Pr[X=i]=\sum_{i=0}^{\infty}\sum_{j=1}^{i}\Pr[X=i]=\sum_{j=1}^{\infty}\sum_{i=j}^{\infty}\Pr[X=i]=\sum_{j=1}^{\infty}\Pr[X\ge j].\quad\square \]
(The key step is swapping the order of summation: instead of summing over \(i\) and counting 1 for each \(j\le i\), sum over \(j\) and count all \(i\ge j\).)
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | (Expected Value) For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):<br>\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i]}} |
<b>(Expected Value)</b> For a random variable \(X\) with \(W_X\subseteq\mathbb{N}_0\):<br>\[ \mathbb{E}[X] = {{c1::\sum_{i=1}^{\infty}\Pr[X\ge i] :: bound form}} \]<br><em>Proof Included</em> |
Note 10: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_4064b971
Before
Front
(Lemma 2.23) Events \(A_1,\ldots,A_n\) are mutually independent iff for all \((s_1,\ldots,s_n)\in\{0,1\}^n\):
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
Back
(Lemma 2.23) Events \(A_1,\ldots,A_n\) are mutually independent iff for all \((s_1,\ldots,s_n)\in\{0,1\}^n\):
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
Proof sketch (⇒): Induction on the number of zeros in \((s_1,\ldots,s_n)\).
Base case (\(s_i=1\) for all \(i\)): direct from definition.
Inductive step (say \(s_1=0\)):
\(\Pr[\bar{A}_1\cap A_2^{s_2}\cap\cdots] = \Pr[A_2^{s_2}\cap\cdots]-\Pr[A_1\cap A_2^{s_2}\cap\cdots]\)
\(= \Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]-\Pr[A_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\)
\(= (1-\Pr[A_1])\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}] = \Pr[\bar{A}_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\). \(\square\)
Base case (\(s_i=1\) for all \(i\)): direct from definition.
Inductive step (say \(s_1=0\)):
\(\Pr[\bar{A}_1\cap A_2^{s_2}\cap\cdots] = \Pr[A_2^{s_2}\cap\cdots]-\Pr[A_1\cap A_2^{s_2}\cap\cdots]\)
\(= \Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]-\Pr[A_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\)
\(= (1-\Pr[A_1])\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}] = \Pr[\bar{A}_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\). \(\square\)
After
Front
Independence Events \(A_1,\ldots,A_n\) are mutually independent iff for all \((s_1,\ldots,s_n)\in\{0,1\}^n\):
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
Back
Independence Events \(A_1,\ldots,A_n\) are mutually independent iff for all \((s_1,\ldots,s_n)\in\{0,1\}^n\):
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]
where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). Proof Included
Proof sketch (⇒): Induction on the number of zeros in \((s_1,\ldots,s_n)\).
Base case (\(s_i=1\) for all \(i\)): direct from definition.
Inductive step (say \(s_1=0\)):
\(\Pr[\bar{A}_1\cap A_2^{s_2}\cap\cdots] = \Pr[A_2^{s_2}\cap\cdots]-\Pr[A_1\cap A_2^{s_2}\cap\cdots]\)
\(= \Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]-\Pr[A_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\)
\(= (1-\Pr[A_1])\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}] = \Pr[\bar{A}_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\). \(\square\)
Base case (\(s_i=1\) for all \(i\)): direct from definition.
Inductive step (say \(s_1=0\)):
\(\Pr[\bar{A}_1\cap A_2^{s_2}\cap\cdots] = \Pr[A_2^{s_2}\cap\cdots]-\Pr[A_1\cap A_2^{s_2}\cap\cdots]\)
\(= \Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]-\Pr[A_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\)
\(= (1-\Pr[A_1])\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}] = \Pr[\bar{A}_1]\Pr[A_2^{s_2}]\cdots\Pr[A_n^{s_n}]\). \(\square\)
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | <b>Independence</b> Events \(A_1,\ldots,A_n\) are <b>mutually</b> independent iff for <b>all</b> \((s_1,\ldots,s_n)\in\{0,1\}^n\):<br>\[ {{c1:: \Pr\!\left[A_1^{s_1}\cap\cdots\cap A_n^{s_n}\right] = \Pr[A_1^{s_1}]\cdots\Pr[A_n^{s_n}]}} \]<br>where \(A_i^1 = A_i\) and \(A_i^0 = \bar{A}_i\). <em>Proof Included</em> |
Note 11: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_55f36256
Before
Front
The distribution function (Verteilungsfunktion) of \(X\) is:
\[ {{c1:: F_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X\le x]}} = \sum_{\substack{x'\in W_X\\x'\le x\Pr[X=x']}}. \]
\[ {{c1:: F_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X\le x]}} = \sum_{\substack{x'\in W_X\\x'\le x\Pr[X=x']}}. \]
Back
The distribution function (Verteilungsfunktion) of \(X\) is:
\[ {{c1:: F_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X\le x]}} = \sum_{\substack{x'\in W_X\\x'\le x\Pr[X=x']}}. \]
\[ {{c1:: F_X : \mathbb{R}\to[0,1], \quad x\mapsto \Pr[X\le x]}} = \sum_{\substack{x'\in W_X\\x'\le x\Pr[X=x']}}. \]
It is non-decreasing, right-continuous, with \(F_X\to 0\) as \(x\to-\infty\) and \(F_X\to 1\) as \(x\to+\infty\).
After
Front
The distribution function (Verteilungsfunktion) of \(X\) is:
\[ {{c1:: F_X : \mathbb{R}\to[0,1], x\mapsto \Pr[X\le x]}} = {{c2::\sum_{\substack{x'\in W_X\\x'\le x} } \Pr[X=x'] }}\]
\[ {{c1:: F_X : \mathbb{R}\to[0,1], x\mapsto \Pr[X\le x]}} = {{c2::\sum_{\substack{x'\in W_X\\x'\le x} } \Pr[X=x'] }}\]
Back
The distribution function (Verteilungsfunktion) of \(X\) is:
\[ {{c1:: F_X : \mathbb{R}\to[0,1], x\mapsto \Pr[X\le x]}} = {{c2::\sum_{\substack{x'\in W_X\\x'\le x} } \Pr[X=x'] }}\]
\[ {{c1:: F_X : \mathbb{R}\to[0,1], x\mapsto \Pr[X\le x]}} = {{c2::\sum_{\substack{x'\in W_X\\x'\le x} } \Pr[X=x'] }}\]
It is non-decreasing, right-continuous, with \(F_X\to 0\) as \(x\to-\infty\) and \(F_X\to 1\) as \(x\to+\infty\).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | The <strong>distribution function</strong> (Verteilungsfunktion) of \(X\) is:<br>\[ {{c1:: F_X : \mathbb{R}\to[0,1], |
The <strong>distribution function</strong> (Verteilungsfunktion) of \(X\) is:<br>\[ {{c1:: F_X : \mathbb{R}\to[0,1], x\mapsto \Pr[X\le x]}} = {{c2::\sum_{\substack{x'\in W_X\\x'\le x} } \Pr[X=x'] }}\] |
Note 12: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Classic
GUID:
modified
Note Type: Horvath Classic
GUID:
sc_59ff89ba
Before
Front
When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) undefined (Def 2.27)?
Back
When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) undefined (Def 2.27)?
The expected value is only defined if the sum converges absolutely, i.e., \(\sum_{x\in W_X}|x|\cdot\Pr[X=x]<\infty\).
If the sum does not converge absolutely (e.g., the positive and negative parts both diverge), \(\mathbb{E}[X]\) is undefined.
For finite probability spaces this is always satisfied (finitely many terms). For infinite spaces, care is needed.
If the sum does not converge absolutely (e.g., the positive and negative parts both diverge), \(\mathbb{E}[X]\) is undefined.
For finite probability spaces this is always satisfied (finitely many terms). For infinite spaces, care is needed.
After
Front
When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) undefined ?
Back
When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) undefined ?
The expected value is only defined if the sum converges absolutely, i.e., \(\sum_{x\in W_X}|x|\cdot\Pr[X=x]<\infty\).
If the sum does not converge absolutely (e.g., the positive and negative parts both diverge), \(\mathbb{E}[X]\) is undefined.
For finite probability spaces this is always satisfied (finitely many terms). For infinite spaces, care is needed.
If the sum does not converge absolutely (e.g., the positive and negative parts both diverge), \(\mathbb{E}[X]\) is undefined.
For finite probability spaces this is always satisfied (finitely many terms). For infinite spaces, care is needed.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Front | When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) <strong>undefined</strong> |
When is the expected value \(\mathbb{E}[X] = \sum_{x\in W_X} x\cdot\Pr[X=x]\) <strong>undefined</strong> ? |
| Back | The expected value is only defined if the sum <strong>converges absolutely</strong>, i.e., \(\sum_{x\in W_X}|x|\cdot\Pr[X=x] |
The expected value is only defined if the sum <strong>converges absolutely</strong>, i.e., \(\sum_{x\in W_X}|x|\cdot\Pr[X=x]<\infty\).<br><br>If the sum does not converge absolutely (e.g., the positive and negative parts both diverge), \(\mathbb{E}[X]\) is <strong>undefined</strong>.<br><br>For <strong>finite</strong> probability spaces this is always satisfied (finitely many terms). For infinite spaces, care is needed. |
Note 13: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Classic
GUID:
modified
Note Type: Horvath Classic
GUID:
sc_65e8ec7d
Before
Front
In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. Why is \(\mathbb{E}[G]\) undefined?
Back
In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. Why is \(\mathbb{E}[G]\) undefined?
\(\Pr[\text{first heads at flip }k] = (1/2)^k\).
The sum in Definition 2.27:
\[ \sum_{k=1}^{\infty}(-1)^{k-1}\cdot 2^k\cdot (1/2)^k = \sum_{k=1}^{\infty}(-1)^{k-1} = +1-1+1-1+\cdots \]
This series does not converge (it oscillates), so \(\mathbb{E}[G]\) is undefined.
Similarly: if the bank always pays \(2^k\), each term equals 1 and the sum diverges to \(+\infty\).
The sum in Definition 2.27:
\[ \sum_{k=1}^{\infty}(-1)^{k-1}\cdot 2^k\cdot (1/2)^k = \sum_{k=1}^{\infty}(-1)^{k-1} = +1-1+1-1+\cdots \]
This series does not converge (it oscillates), so \(\mathbb{E}[G]\) is undefined.
Similarly: if the bank always pays \(2^k\), each term equals 1 and the sum diverges to \(+\infty\).
After
Front
In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. What is \(\mathbb{E}[G]\)?
Back
In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. What is \(\mathbb{E}[G]\)?
\(\Pr[\text{first heads at flip }k] = (1/2)^k\).
The sum in Definition 2.27:
\[ \sum_{k=1}^{\infty}(-1)^{k-1}\cdot 2^k\cdot (1/2)^k = \sum_{k=1}^{\infty}(-1)^{k-1} = +1-1+1-1+\cdots \]
This series does not converge (it oscillates), so \(\mathbb{E}[G]\) is undefined.
Similarly: if the bank always pays \(2^k\), each term equals 1 and the sum diverges to \(+\infty\).
The sum in Definition 2.27:
\[ \sum_{k=1}^{\infty}(-1)^{k-1}\cdot 2^k\cdot (1/2)^k = \sum_{k=1}^{\infty}(-1)^{k-1} = +1-1+1-1+\cdots \]
This series does not converge (it oscillates), so \(\mathbb{E}[G]\) is undefined.
Similarly: if the bank always pays \(2^k\), each term equals 1 and the sum diverges to \(+\infty\).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Front | In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. Wh |
In the casino game: flip a coin until the first heads after \(k\) flips. Bank gains \(2^k\) if \(k\) is odd, loses \(2^k\) if \(k\) is even. What is \(\mathbb{E}[G]\)? |
Note 14: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_79bd603a
Before
Front
If \(A\) and \(B\) are independent, prove that \(\bar{A}\) and \(B\) are also independent. Proof Included
Back
If \(A\) and \(B\) are independent, prove that \(\bar{A}\) and \(B\) are also independent. Proof Included
Need: \(\Pr[\bar{A}\cap B]=\Pr[\bar{A}]\cdot\Pr[B]\).
\[ \Pr[\bar{A}\cap B] = \Pr[B] - \Pr[A\cap B] = \Pr[B] - \Pr[A]\Pr[B] = (1-\Pr[A])\Pr[B] = \Pr[\bar{A}]\Pr[B]. \quad\square \]
Consequence: If \(A_1,\ldots,A_n\) are mutually independent, so is any family obtained by replacing some \(A_i\) with \(\bar{A}_i\) (Lemma 2.23).
\[ \Pr[\bar{A}\cap B] = \Pr[B] - \Pr[A\cap B] = \Pr[B] - \Pr[A]\Pr[B] = (1-\Pr[A])\Pr[B] = \Pr[\bar{A}]\Pr[B]. \quad\square \]
Consequence: If \(A_1,\ldots,A_n\) are mutually independent, so is any family obtained by replacing some \(A_i\) with \(\bar{A}_i\) (Lemma 2.23).
After
Front
If \(A\) and \(B\) are independent, prove that {{c1:: \(\bar{A}\) and \(B\)::complement}} are also independent. Proof Included
Back
If \(A\) and \(B\) are independent, prove that {{c1:: \(\bar{A}\) and \(B\)::complement}} are also independent. Proof Included
Need: \(\Pr[\bar{A}\cap B]=\Pr[\bar{A}]\cdot\Pr[B]\).
\[ \Pr[\bar{A}\cap B] = \Pr[B] - \Pr[A\cap B] = \Pr[B] - \Pr[A]\Pr[B] = (1-\Pr[A])\Pr[B] = \Pr[\bar{A}]\Pr[B]. \quad\square \]
Consequence: If \(A_1,\ldots,A_n\) are mutually independent, so is any family obtained by replacing some \(A_i\) with \(\bar{A}_i\) (Lemma 2.23).
\[ \Pr[\bar{A}\cap B] = \Pr[B] - \Pr[A\cap B] = \Pr[B] - \Pr[A]\Pr[B] = (1-\Pr[A])\Pr[B] = \Pr[\bar{A}]\Pr[B]. \quad\square \]
Consequence: If \(A_1,\ldots,A_n\) are mutually independent, so is any family obtained by replacing some \(A_i\) with \(\bar{A}_i\) (Lemma 2.23).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | If \(A\) and \(B\) are independent, prove that \(\bar{A}\) and \(B\) are also independent. <em>Proof Included</em> | If \(A\) and \(B\) are independent, prove that {{c1:: \(\bar{A}\) and \(B\)::complement}} are also independent. <em>Proof Included</em> |
Note 15: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Classic
GUID:
modified
Note Type: Horvath Classic
GUID:
sc_9157c9d7
Before
Front
How can linearity of expectation and indicator variables prove the inclusion-exclusion formula (Satz 2.5)? Proof Included
Back
How can linearity of expectation and indicator variables prove the inclusion-exclusion formula (Satz 2.5)? Proof Included
Let \(B=A_1\cup\cdots\cup A_n\). We want \(\Pr[B]=1-\Pr[\bar{B}]\).
Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).
Expand the product:
\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1
Apply \(\mathbb{E}[\cdot]\) and use:
\(\mathbb{E}[I_{A_i}]=\Pr[A_i]\)
\(\mathbb{E}[I_{A_{i_1}}\cdots I_{A_{i_k}}]=\Pr[A_{i_1}\cap\cdots\cap A_{i_k}]\) (product of indicators = indicator of intersection)
Taking expectations gives exactly the inclusion-exclusion formula for \(\Pr[\bar{B}]=1-\Pr[B]\). \(\square\)
Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).
Expand the product:
\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1
Apply \(\mathbb{E}[\cdot]\) and use:
\(\mathbb{E}[I_{A_i}]=\Pr[A_i]\)
\(\mathbb{E}[I_{A_{i_1}}\cdots I_{A_{i_k}}]=\Pr[A_{i_1}\cap\cdots\cap A_{i_k}]\) (product of indicators = indicator of intersection)
Taking expectations gives exactly the inclusion-exclusion formula for \(\Pr[\bar{B}]=1-\Pr[B]\). \(\square\)
After
Front
How can linearity of expectation and indicator variables prove the inclusion-exclusion formula (Satz 2.5)? Proof Included
Back
How can linearity of expectation and indicator variables prove the inclusion-exclusion formula (Satz 2.5)? Proof Included
Let \(B=A_1\cup\cdots\cup A_n\). We want \(\Pr[B]=1-\Pr[\bar{B}]\).
Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).
Expand the product:
\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1<i_2}i_{a_{i_1}}i_{a_{i_2}}-\cdots+(-1)^ni_{a_1}\cdots i_{a_n}\]Apply \(\mathbb{E}[\cdot]\) and use:
\(\mathbb{E}[I_{A_i}]=\Pr[A_i]\)
\(\mathbb{E}[I_{A_{i_1}}\cdots I_{A_{i_k}}]=\Pr[A_{i_1}\cap\cdots\cap A_{i_k}]\) (product of indicators = indicator of intersection)
Taking expectations gives exactly the inclusion-exclusion formula for \(\Pr[\bar{B}]=1-\Pr[B]\). \(\square\)
Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).
Expand the product:
\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1<i_2}i_{a_{i_1}}i_{a_{i_2}}-\cdots+(-1)^ni_{a_1}\cdots i_{a_n}\]Apply \(\mathbb{E}[\cdot]\) and use:
\(\mathbb{E}[I_{A_i}]=\Pr[A_i]\)
\(\mathbb{E}[I_{A_{i_1}}\cdots I_{A_{i_k}}]=\Pr[A_{i_1}\cap\cdots\cap A_{i_k}]\) (product of indicators = indicator of intersection)
Taking expectations gives exactly the inclusion-exclusion formula for \(\Pr[\bar{B}]=1-\Pr[B]\). \(\square\)
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Back | Let \(B=A_1\cup\cdots\cup A_n\). We want \(\Pr[B]=1-\Pr[\bar{B}]\).<br><br>Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).<br><br>Expand the product:<br>\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1 |
Let \(B=A_1\cup\cdots\cup A_n\). We want \(\Pr[B]=1-\Pr[\bar{B}]\).<br><br>Note: \(I_{\bar{B}}=\prod_{i=1}^n(1-I_{A_i})\) (since \(\bar{B}\) occurs iff all \(A_i\) fail).<br><br>Expand the product:<br>\[ I_{\bar{B}}=1-\sum_{i}I_{A_i}+\sum_{i_1<i_2}i_{a_{i_1}}i_{a_{i_2}}-\cdots+(-1)^ni_{a_1}\cdots i_{a_n}\]Apply \(\mathbb{E}[\cdot]\) and use:<br>\(\mathbb{E}[I_{A_i}]=\Pr[A_i]\)<br>\(\mathbb{E}[I_{A_{i_1}}\cdots I_{A_{i_k}}]=\Pr[A_{i_1}\cap\cdots\cap A_{i_k}]\) (product of indicators = indicator of intersection)<br><br>Taking expectations gives exactly the inclusion-exclusion formula for \(\Pr[\bar{B}]=1-\Pr[B]\). \(\square\) |
Note 16: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_a7abceb1
Before
Front
(Satz von der totalen Wahrscheinlichkeit) Let \(A_1,\ldots,A_n\) be pairwise disjoint with \(B\subseteq A_1\cup\cdots\cup A_n\). Then:
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Back
(Satz von der totalen Wahrscheinlichkeit) Let \(A_1,\ldots,A_n\) be pairwise disjoint with \(B\subseteq A_1\cup\cdots\cup A_n\). Then:
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Proof:
Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).
Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\).
Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)
Use: Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe).
Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).
Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\).
Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)
Use: Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe).
After
Front
(Satz von der totalen Wahrscheinlichkeit) Let \(A_1,\ldots,A_n\) be pairwise disjoint with \(B\subseteq A_1\cup\cdots\cup A_n\). Then:
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Back
(Satz von der totalen Wahrscheinlichkeit) Let \(A_1,\ldots,A_n\) be pairwise disjoint with \(B\subseteq A_1\cup\cdots\cup A_n\). Then:
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
\[ \Pr[B] = {{c1::\sum_{i=1}^{n}\Pr[B|A_i]\cdot\Pr[A_i]}}. \]
Proof Included
Proof:
Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).
Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\). (as disjoint)
Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)
Use: Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe).
Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).
Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\). (as disjoint)
Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)
Use: Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe).
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Extra | <strong>Proof:</strong><br>Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).<br>Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\).<br>Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)<br><br><strong>Use:</strong> Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe). | <strong>Proof:</strong><br>Decompose: \(B = (B\cap A_1)\cup\cdots\cup(B\cap A_n)\) (disjoint parts).<br>Apply Additionssatz: \(\Pr[B] = \sum_i \Pr[B\cap A_i]\). (as disjoint)<br>Use \(\Pr[B\cap A_i] = \Pr[B|A_i]\cdot\Pr[A_i]\) (from definition of conditional probability). \(\square\)<br><br><strong>Use:</strong> Decompose a complex event into simpler cases (the \(A_i\) form a partition of the relevant universe). |
Note 17: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_d6976913
Before
Front
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable (Indikatorvariable) \(X_A\) is defined by:
\[ X_A(\omega) := \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} \]
\(X_A\) takes only the values 0 and 1, and its expected value is \(\mathbb{E}[X_A] = \Pr[A]\).
\[ X_A(\omega) := \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} \]
\(X_A\) takes only the values 0 and 1, and its expected value is \(\mathbb{E}[X_A] = \Pr[A]\).
Back
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable (Indikatorvariable) \(X_A\) is defined by:
\[ X_A(\omega) := \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} \]
\(X_A\) takes only the values 0 and 1, and its expected value is \(\mathbb{E}[X_A] = \Pr[A]\).
\[ X_A(\omega) := \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} \]
\(X_A\) takes only the values 0 and 1, and its expected value is \(\mathbb{E}[X_A] = \Pr[A]\).
Indicator variables are also called Bernoulli variables. They are the bridge between events and random variables: the probability of an event equals the expected value of its indicator.
Every indicator \(X_A\) is a \(\text{Bernoulli}(\Pr[A])\) random variable. The name "indicator" reflects the fact that \(X_A(\omega)=1\) precisely when \(\omega\) "indicates" that \(A\) has occurred.
Every indicator \(X_A\) is a \(\text{Bernoulli}(\Pr[A])\) random variable. The name "indicator" reflects the fact that \(X_A(\omega)=1\) precisely when \(\omega\) "indicates" that \(A\) has occurred.
After
Front
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable (Indikatorvariable) \(X_A\) is defined by:
\[ X_A(\omega) := {{c1:: \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} }}\]
\[ X_A(\omega) := {{c1:: \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} }}\]
Back
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable (Indikatorvariable) \(X_A\) is defined by:
\[ X_A(\omega) := {{c1:: \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} }}\]
\[ X_A(\omega) := {{c1:: \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} }}\]
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | (Beobachtung 2.35) For an event \(A\subseteq\Omega\), the <strong>indicator variable</strong> (Indikatorvariable) \(X_A\) is defined by:<br>\[ X_A(\omega) := \begin{cases} |
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the <strong>indicator variable</strong> (Indikatorvariable) \(X_A\) is defined by:<br>\[ X_A(\omega) := {{c1:: \begin{cases} 1 & \text{falls } \omega \in A, \\ 0 & \text{sonst.} \end{cases} }}\]<br> |
| Extra |
Note 18: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_d8c88bc2
Before
Front
For a random variable \(X:\Omega\to\mathbb{R}\), the range (Wertebereich) is:
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\}}}. \]
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\}}}. \]
Back
For a random variable \(X:\Omega\to\mathbb{R}\), the range (Wertebereich) is:
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\}}}. \]
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\}}}. \]
After
Front
For a random variable \(X:\Omega\to\mathbb{R}\), the range (Wertebereich) is:
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\} }}\]
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\} }}\]
Back
For a random variable \(X:\Omega\to\mathbb{R}\), the range (Wertebereich) is:
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\} }}\]
\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\} }}\]
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | For a random variable \(X:\Omega\to\mathbb{R}\), the <strong>range</strong> (Wertebereich) is:<br>\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\}}} |
For a random variable \(X:\Omega\to\mathbb{R}\), the <strong>range</strong> (Wertebereich) is:<br>\[ W_X := {{c1::X(\Omega) = \{x\in\mathbb{R}\mid\exists\omega\in\Omega:\, X(\omega)=x\} }}\] |
Note 19: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_deaab462
Before
Front
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable \(X_A\) satisfies:
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
Back
(Beobachtung 2.35) For an event \(A\subseteq\Omega\), the indicator variable \(X_A\) satisfies:
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
Proof: \(\mathbb{E}[X_A]=1\cdot\Pr[X_A=1]+0\cdot\Pr[X_A=0]=\Pr[A].\quad\square\)
This is the bridge between events (probability) and random variables (expectation): an event's probability equals the expected value of its indicator.
This is the bridge between events (probability) and random variables (expectation): an event's probability equals the expected value of its indicator.
After
Front
For an event \(A\subseteq\Omega\), the indicator variable \(X_A\) satisfies:
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
Back
For an event \(A\subseteq\Omega\), the indicator variable \(X_A\) satisfies:
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
\[ \mathbb{E}[X_A] = \Pr[A]. \]
Proof Included
Proof: \(\mathbb{E}[X_A]=1\cdot\Pr[X_A=1]+0\cdot\Pr[X_A=0]=\Pr[A].\quad\square\)
This is the bridge between events (probability) and random variables (expectation): an event's probability equals the expected value of its indicator.
This is the bridge between events (probability) and random variables (expectation): an event's probability equals the expected value of its indicator.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | For an event \(A\subseteq\Omega\), the indicator variable \(X_A\) satisfies:<br>\[ \mathbb{E}[X_A] = {{c1::\Pr[A]}}. \]<br><em>Proof Included</em> |
Note 20: ETH::2. Semester::A&W
Deck: ETH::2. Semester::A&W
Note Type: Horvath Cloze
GUID:
modified
Note Type: Horvath Cloze
GUID:
sc_ee1aa72c
Before
Front
Since \(\Pr[\cdot|B]\) is itself a valid probability space, the following hold:
\(\Pr[\bar{A}|B] = 1 - \Pr[A|B]\)
\(\Pr[\emptyset|B] = 0\)
If \(A\subseteq C\) then \(\Pr[A|B] \le \Pr[C|B]\)
\(\Pr[\bar{A}|B] = 1 - \Pr[A|B]\)
\(\Pr[\emptyset|B] = 0\)
If \(A\subseteq C\) then \(\Pr[A|B] \le \Pr[C|B]\)
Back
Since \(\Pr[\cdot|B]\) is itself a valid probability space, the following hold:
\(\Pr[\bar{A}|B] = 1 - \Pr[A|B]\)
\(\Pr[\emptyset|B] = 0\)
If \(A\subseteq C\) then \(\Pr[A|B] \le \Pr[C|B]\)
\(\Pr[\bar{A}|B] = 1 - \Pr[A|B]\)
\(\Pr[\emptyset|B] = 0\)
If \(A\subseteq C\) then \(\Pr[A|B] \le \Pr[C|B]\)
This is because once we fix the conditioning event \(B\), all standard probability axioms apply within the conditional space.
After
Front
Since \(\Pr[\cdot|B]\) is itself a valid probability space.
Back
Since \(\Pr[\cdot|B]\) is itself a valid probability space.
This is because once we fix the conditioning event \(B\), all standard probability axioms apply within the conditional space.
Field-by-field Comparison
| Field | Before | After |
|---|---|---|
| Text | Since \(\Pr[\cdot|B]\) is itself a valid probability space |
Since \(\Pr[\cdot|B]\) is itself a {{c1::valid probability space}}. |