Math Basis of Statistic Learning
好像很久没有写了呢,看了percy Liang的教学笔记,没有看完但是觉得很有收获的,教学内容被percy Liang分成了4个部分, 从asymptotics到uniform convergence到online learning外加一个kernel method。在看的时候其实是有非常多的东西是看不懂的, 尤其是统计学习理论有非常多的数学基础,所幸的是appendix中的结论简洁明了,不绕圈子,所以很想总结一下,估计以后一定会用到的。
In the following, assume the following:
Let be real vectors drawn i.i.d. from some distribution with mean and covariance matrix .
Convergence in probability
We say that a sequence of random variables () converges in probability to a random variable (written )
if for all , we have Example:
(weak law of large numbers)
Convergence in distribution
We say that a sequence of distributions () converges in distribution (weak convergence) to a distribution (written ) if for all bounded continuous functions ,
When we write &&Y_{n} \overset{D}{\rightarrow} Y&& for a sequence of random variables (), we mean that the sequence of distribution of those random variables converges in distribution.
Example:
(central limit theorem)
Continuous mapping theorem
This theorem allows us to transfer convergence in probability and convergence in distribution results through continuous transformations.
If is continuous and , then
If is continuous and &&Y_{n} \overset{D}{\rightarrow} Y&&, then
Example:
Delta method
This theorem allows us to transfer asymptotic normality results through smooth transformations.
If is continous, and , then
The intuition is Talor approximation: