Processing math: 100%

Convergence of Random Variables


[1] Convergence in probability
Definition:  A sequence of random variables X_{1}, X_{2},... converges in probability to random variable X if for every e > 0, \lim_{n\rightarrow \infty} P(|X_{n}-X| > e )=0 \Leftrightarrow \lim_{n\rightarrow \infty} P(|X_{n}-X| \leq e )=1

Think about a sequence of random variables. This sequence asymptotically reaches a certain random variable X (or a constant). How do we know it? We can define a boundary of the certain random variable. The sequence will reach this boundary, however we never know it actually reach the certain random variable X.


 
[2] Almost Sure Convergence

Definition: A sequence of random variables X_{1}, X_{2},... almost surely to a random variable if for every e>0, P(\lim_{n\rightarrow \infty}|X_{n}-X| \leq e )=1  
A sequence of random variables definitely reach a certain random variable X. But we don’t know when it actually reach it. Therefore, after a certain n, then the sequence of random variables are equal to the variable X.
 
[3] Convergence in distribution.
Definition: A sequence of random variables X_{1}, X_{2},... converge in distribution to a random variable X if \lim_{n\rightarrow \infty}P(|X_{n} \leq x)= P(X\leq x)=F_{X}x at all points x where F(x) is continuous.
 

* Central Limit Theorem 
Definition: Suppose that X_{1}, X_{2},..., X_{n} are i.i.d. random variable with mean \mu and variance \sigma^2. Then, Z_{n}=\frac{\sqrt{n}(\bar{X}-\mu)}{\sigma}\rightarrow Z\sim N(0,1) (in distribution)  

Remark) The condition of the CLT is different from the others. There should be i.i.d, number of n randm varaibles.  

Varance stabilizing transformations
- When g is differentiable, we have \sqrt{n}(g(\bar{X_{n}}-g(\mu)))\rightarrow g'(\mu)N(0, \sigma^2) in distribution.  
 
 

No comments:

Post a Comment