Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[lln_clt] Update editorial suggestions #415

Merged
merged 2 commits into from
Apr 9, 2024
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 5 additions & 5 deletions lectures/lln_clt.md
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ print(X)
```

In this setting, the LLN tells us if we flip the coin many times, the fraction
of heads that we see will be close to the mean $p$.
of heads that we see will be close to the mean $p$. We use $n$ to represent the number of times the coin is flipped.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@SylviaZhaooo Would you mind to put the second sentence on a separate line -- this is our (unusual) convention for the lectures.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure!


Let's check this:

Expand Down Expand Up @@ -286,7 +286,7 @@ as expected.

Let's vary `n` to see how the distribution of the sample mean changes.

We will use a violin plot to show the different distributions.
We will use a [violin plot](https://intro.quantecon.org/prob_dist.html#violin-plots) to show the different distributions.

Each distribution in the violin plot represents the distribution of $X_n$ for some $n$, calculated by simulation.

Expand Down Expand Up @@ -357,7 +357,7 @@ This means that the distribution of $\bar X_n$ does not eventually concentrate o

Hence the LLN does not hold.

The LLN fails to hold here because the assumption $\mathbb E|X| = \infty$ is violated by the Cauchy distribution.
The LLN fails to hold here because the assumption $\mathbb E|X| < \infty$ is violated by the Cauchy distribution.

+++

Expand Down Expand Up @@ -438,7 +438,7 @@ Here $\stackrel { d } {\to} N(0, \sigma^2)$ indicates [convergence in distributi

The striking implication of the CLT is that for **any** distribution with
finite [second moment](https://en.wikipedia.org/wiki/Moment_(mathematics)), the simple operation of adding independent
copies **always** leads to a Gaussian curve.
copies **always** leads to a Gaussian(Normal) curve.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
copies **always** leads to a Gaussian(Normal) curve.
copies **always** leads to a Gaussian (Normal) curve.

This is a really minor point but @jstac re: style I think there should be space between Gaussian and (Normal). Do you agree?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes please, thanks @mmcky and @SylviaZhaooo




Expand Down Expand Up @@ -503,7 +503,7 @@ The fit to the normal density is already tight and can be further improved by in
```{exercise}
:label: lln_ex1

Repeat the simulation [above1](sim_one) with the [Beta distribution](https://en.wikipedia.org/wiki/Beta_distribution).
Repeat the simulation [above](sim_one) with the [Beta distribution](https://en.wikipedia.org/wiki/Beta_distribution).

You can choose any $\alpha > 0$ and $\beta > 0$.
```
Expand Down
Loading