[1]
Convergence in probability
Definition: A sequence of random variables X_{1}, X_{2},... converges in probability to random variable X if for every e > 0, \lim_{n\rightarrow \infty} P(|X_{n}-X| > e )=0 \Leftrightarrow \lim_{n\rightarrow \infty} P(|X_{n}-X| \leq e )=1Think about a sequence of random variables. This sequence asymptotically reaches a certain random variable X (or a constant). How do we know it? We can define a boundary of the certain random variable. The sequence will reach this boundary, however we never know it actually reach the certain random variable X.
[2] Almost Sure Convergence
Definition: A sequence of random variables X_{1}, X_{2},... almost surely to a random variable if for every e>0, P(\lim_{n\rightarrow \infty}|X_{n}-X| \leq e )=1
A
sequence of random variables definitely reach a certain random variable X. But
we don’t know when it actually reach it. Therefore, after a certain n, then the
sequence of random variables are equal to the variable X.
[3]
Convergence in distribution.
Definition: A sequence of random variables X_{1}, X_{2},... converge in distribution to a random variable X if \lim_{n\rightarrow \infty}P(|X_{n} \leq x)= P(X\leq x)=F_{X}x at all points x where F(x) is continuous.
*
Central Limit Theorem
Definition: Suppose that X_{1}, X_{2},..., X_{n} are i.i.d. random variable with mean \mu and variance \sigma^2. Then, Z_{n}=\frac{\sqrt{n}(\bar{X}-\mu)}{\sigma}\rightarrow Z\sim N(0,1) (in distribution) Remark) The condition of the CLT is different from the others. There should be i.i.d, number of n randm varaibles.
Varance stabilizing transformations
- When g is differentiable, we have \sqrt{n}(g(\bar{X_{n}}-g(\mu)))\rightarrow g'(\mu)N(0, \sigma^2) in distribution.
No comments:
Post a Comment