Math Basis of Statistic Learning

2 March 2017 - Beijing

好像很久没有写了呢,看了percy Liang的教学笔记,没有看完但是觉得很有收获的,教学内容被percy Liang分成了4个部分, 从asymptotics到uniform convergence到online learning外加一个kernel method。在看的时候其实是有非常多的东西是看不懂的, 尤其是统计学习理论有非常多的数学基础,所幸的是appendix中的结论简洁明了,不绕圈子,所以很想总结一下,估计以后一定会用到的。

In the following, assume the following:

Let be real vectors drawn i.i.d. from some distribution with mean and covariance matrix .

Convergence in probability

We say that a sequence of random variables () converges in probability to a random variable (written )

if for all , we have Example:

(weak law of large numbers)

Convergence in distribution

We say that a sequence of distributions () converges in distribution (weak convergence) to a distribution (written ) if for all bounded continuous functions ,

When we write &&Y_{n} \overset{D}{\rightarrow} Y&& for a sequence of random variables (), we mean that the sequence of distribution of those random variables converges in distribution.

Example:

(central limit theorem)

Continuous mapping theorem

This theorem allows us to transfer convergence in probability and convergence in distribution results through continuous transformations.

If is continuous and , then

If is continuous and &&Y_{n} \overset{D}{\rightarrow} Y&&, then

Example:

Delta method

This theorem allows us to transfer asymptotic normality results through smooth transformations.

If is continous, and , then

The intuition is Talor approximation:

Written on March 2, 2017