Skip to content

Commit 73f7b5a

Browse files
Update README.md
1 parent 8116e8a commit 73f7b5a

File tree

1 file changed

+30
-2
lines changed

1 file changed

+30
-2
lines changed

README.md

+30-2
Original file line numberDiff line numberDiff line change
@@ -41,8 +41,8 @@ Information Gain-
4141

4242
Criteria to measure the impurity of a node I(node):
4343
1. Variance (Regression) [Variance reduction of a node N is defined as the total reduction of the variance of the target variable x due to the split at this node]
44-
2. Gini impurity (Classification) [Gini impurity is a measure of how often a randomly chosen element from the set would be incorrectly labeled if it was randomly labeled according to the distribution of labels in the subset]
45-
3. Entropy (Classification) [Information entropy is the average rate at which information is produced by a stochastic source of data]
44+
2. Gini impurity (Classification) [Measure of impurity. Gini impurity is a measure of how often a randomly chosen element from the set would be incorrectly labeled if it was randomly labeled according to the distribution of labels in the subset]
45+
3. Entropy (Classification) [Measure of purity. Information entropy is the average rate at which information is produced by a stochastic source of data]
4646

4747
Note
4848
- Most of the time, the gini index and entropy lead to the same results.
@@ -191,3 +191,31 @@ Feature Importance
191191
- In sklearn :
192192
- how much the tree nodes use a particular feature (weighted average) to reduce impurity
193193
- accessed using the attribute feature_importance_
194+
195+
196+
Part 4 BOOSTING
197+
198+
- Boosting refers to an ensemble method in which several models are trained sequentially with each model learning from the errors of its predecessors.
199+
- Boosting: Ensemble method combining several weak learners to form a strong learner.
200+
- Weak learner: Model doing slightly better than random guessing.
201+
- Example of weak learner: Decision stump (CART whose maximum depth is 1).
202+
- Train an ensemble of predictors sequentially.
203+
- Each predictor tries to correct its predecessor.
204+
- Most popular boosting methods: AdaBoost, Gradient Boosting.
205+
206+
Adaboost
207+
- Stands for Adaptive Boosting.
208+
- Each predictor pays more attention to the instances wrongly predicted by its predecessor.
209+
- Achieved by changing the weights of training instances.
210+
- Each predictor is assigned a coefficient α.
211+
- α depends on the predictor's training error
212+
- Learning rate: 0 < η ≤ 1. It help to shrink coeeficient α. It is the tradeoff between η and number of estimator.
213+
- Smaller number of η should be compensiated by high number of estimator.
214+
215+
AdaBoost Classication:
216+
- Weighted majority voting.
217+
- In sklearn: AdaBoostClassifier.
218+
219+
AdaBoost Regression:
220+
- Weighted average.
221+
- In sklearn: AdaBoostRegressor.

0 commit comments

Comments
 (0)