您的位置:首页 > 其它

Note for video Machine Learning and Data Mining——error and noise

2014-10-05 20:10 459 查看
Here is the note for lecture four.

error measures

When it comes to final hypothesis, we hope H(hypothesis) approximate f(target) as well as possible.

It can be written as

H ≈ f

Error measures could be defined as

E = 1/n ∑(f(xi)-H(xi))

If xi belongs to x1...xn, that means xi is one of the sample. Then E => Ein, which means sample's error measures.

And there will be a variable Eout, which means ensemble's error measures.

Generally, We hope Eout to be minimize, approximate zero. But it is impossible to get a real Eout.To make hypothesis better,

firstly, we could minimize Ein. And then make Ein approximate Eout, that's OK.

However, there will be a problem. If the model becomes complex to minimize Ein, Ein tracks Eout will much more loosely than it used to.

So remember two points,

1. Minimize Ein

2. Ein is close enough to Eout

Noisy Targets

In reality, when it comes to a problem, identical conditions sometimes have different results. This leads to Noise.

Because of this problem, there will be some change in target function.
So instead of target function y=f(x), we use

target distribution

P (y|x)

Here most of y is possible and rest is impossible. So we can define noise with y - E(y|x).
内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: