首页 > 代码库 > 机器学习笔记(Washington University)- Classification Specialization-week five
机器学习笔记(Washington University)- Classification Specialization-week five
1. Ensemble classifier
Each classifier votes on prediction
Ensemble model = sign(w1f1(xi) + w2f2(xi) + w3f3(xi))
w1 w2 w3 is the learning coefficients
f1(xi), f2(xi), f3(xi)) is three classifiers
2. Boosting
Focus on hard or more important pointsand keep adding new classfier.
Boosting is more robust to overfitting but we still need carefully to choose boosting captical T
using validation set or cross validation.
3. Adaboost
1. Start with weight for all points: αi = 1/N
For t = 1 ... T
- Learn ft(x) with data weights αi
- Compute coefficient wt
- Note :
Adaboost use the formual below to compute coefficient wt of classifier ft(x)
wt = 1/2*ln(1- weighted_error(ft)/weighted_error(ft))
- Note :
- Recompute weights αi
- αi = αie-Wt, if ft(xi)=yi else αieWt
- Normalizing weights:
- αi = αi / (α1 +α2 ... αN)
Final model predicts the value by:
y = sign(w1 f1(x) + wt ft(x) ... wT fT(x))
Weighted classification error:
weighted_error = total weight of mistakes / total weights of all data points
Normalizing weights αi
normalize weights to add up to 1 after every iterationn
αi = αi / (α1 +α2 ... αN)
4. Adaboost Theorem
if we can find a weak leatner with weighted_error < 0.5 (beat random guess) at every iteration t,
the training error of boosted classifier goes to zero as the iterations of boosting goes to infinity.
机器学习笔记(Washington University)- Classification Specialization-week five