Note for video Machine Learning and Data Mining——error and noise
2014-10-05 20:10
459 查看
Here is the note for lecture four.
error measures
When it comes to final hypothesis, we hope H(hypothesis) approximate f(target) as well as possible.
It can be written as
H ≈ f
Error measures could be defined as
E = 1/n ∑(f(xi)-H(xi))
If xi belongs to x1...xn, that means xi is one of the sample. Then E => Ein, which means sample's error measures.
And there will be a variable Eout, which means ensemble's error measures.
Generally, We hope Eout to be minimize, approximate zero. But it is impossible to get a real Eout.To make hypothesis better,
firstly, we could minimize Ein. And then make Ein approximate Eout, that's OK.
However, there will be a problem. If the model becomes complex to minimize Ein, Ein tracks Eout will much more loosely than it used to.
So remember two points,
1. Minimize Ein
2. Ein is close enough to Eout
Noisy Targets
In reality, when it comes to a problem, identical conditions sometimes have different results. This leads to Noise.
Because of this problem, there will be some change in target function.
So instead of target function y=f(x), we use
target distribution
P (y|x)
Here most of y is possible and rest is impossible. So we can define noise with y - E(y|x).
error measures
When it comes to final hypothesis, we hope H(hypothesis) approximate f(target) as well as possible.
It can be written as
H ≈ f
Error measures could be defined as
E = 1/n ∑(f(xi)-H(xi))
If xi belongs to x1...xn, that means xi is one of the sample. Then E => Ein, which means sample's error measures.
And there will be a variable Eout, which means ensemble's error measures.
Generally, We hope Eout to be minimize, approximate zero. But it is impossible to get a real Eout.To make hypothesis better,
firstly, we could minimize Ein. And then make Ein approximate Eout, that's OK.
However, there will be a problem. If the model becomes complex to minimize Ein, Ein tracks Eout will much more loosely than it used to.
So remember two points,
1. Minimize Ein
2. Ein is close enough to Eout
Noisy Targets
In reality, when it comes to a problem, identical conditions sometimes have different results. This leads to Noise.
Because of this problem, there will be some change in target function.
So instead of target function y=f(x), we use
target distribution
P (y|x)
Here most of y is possible and rest is impossible. So we can define noise with y - E(y|x).
相关文章推荐
- Note for video Machine Learning and Data Mining——The Learning Problem
- Note for video Machine Learning and Data Mining——training vs Testing
- Note for video Machine Learning and Data Mining——Theory of Generalization
- Note for video Machine Learning and Data Mining——Linear Model
- Note for video Machine Learning and Data Mining——learning is feasible
- Note for video Machine Learning and Data Mining——Linear Model
- Note for video Machine Learning and Data Mining——training vs Testing
- Machine Learning and Data Mining for Computer Security: Methods and Applications
- Machine Learning and Data Mining Course/Video
- CalTech machine learning, video 4(Error & Noise) note
- What’s the difference between machine learning, statistics, and data mining?
- Data Mining and Machine Learning in Cybersecurity PDF
- Machine learning and data mining
- 7 Steps for Learning Data Mining and Data Science
- Machine Learning and Data Mining——2. 机器学习学习路线与资料
- Machine Learning and Data Mining(机器学习与数据挖掘)
- Principles of Data Mining (Adaptive Computation and Machine Learning)
- Machine learning and Data Mining - Association Analysis with Python
- How do you explain Machine Learning and Data Mining to non Computer Science people?
- What is the difference between data mining and machine learning?