Bagging ensemble
1
Background knowledge
 Bias(偏差):describe the difference(or distance) between
predicted result and real result。
 Variance(變異):describe the instability of predicted result。
2
準是 bias 小,確是 variance 小,準確是相對概念,因為
bias-variance tradeoff。
Background knowledge
 Bias-Variance Tradeoff(偏差和變異之權衡)
 Underfitting:when model’s predicted result with high bias, the error of training
set and validation set are very high.
 Overfitting: when model’s predicted result with high variance
the error of training set is very low but validation set is very high.
3
How to deal with?
 How to deal with underfitting:
1. Increasing model iteration times.
2. Using more powerful model to replace.
3. Generate more feature to use.
4. Decreasing regularization.
 How to deal with underfitting:
1. Increasing training set.
2. Decreasing training feature.
3. Increasing regularization.
4
Conception
 Bagging ensemble is first using bootstrap to select training subset randomly
secondly using those training subset to build classifiers, finally using majority
vote to get final result.
Conception
1. Mainly to reduce Variance,have a little effect on Bias
2. Models for High Variance & Low Bias

Bagging ensemble

  • 1.
  • 2.
    Background knowledge  Bias(偏差):describethe difference(or distance) between predicted result and real result。  Variance(變異):describe the instability of predicted result。 2 準是 bias 小,確是 variance 小,準確是相對概念,因為 bias-variance tradeoff。
  • 3.
    Background knowledge  Bias-VarianceTradeoff(偏差和變異之權衡)  Underfitting:when model’s predicted result with high bias, the error of training set and validation set are very high.  Overfitting: when model’s predicted result with high variance the error of training set is very low but validation set is very high. 3
  • 4.
    How to dealwith?  How to deal with underfitting: 1. Increasing model iteration times. 2. Using more powerful model to replace. 3. Generate more feature to use. 4. Decreasing regularization.  How to deal with underfitting: 1. Increasing training set. 2. Decreasing training feature. 3. Increasing regularization. 4
  • 5.
    Conception  Bagging ensembleis first using bootstrap to select training subset randomly secondly using those training subset to build classifiers, finally using majority vote to get final result.
  • 6.
    Conception 1. Mainly toreduce Variance,have a little effect on Bias 2. Models for High Variance & Low Bias

Editor's Notes

  • #7 如果是用Bagging會希望單一分類器能夠是一個效能比較好的分類器。 Bagging的優點在於原始訓練樣本中有噪聲資料(不好的資料),透過Bagging抽樣就有機會不讓有噪聲資料被訓練到,所以可以降低模型的不穩定性。