- Using hellinger distance to split supervised datasets, instead of gini and entropy. Claims better results.
- Visualize decision trees, forests
explains about the similarities and how to measure. which is the best split? based on SSE and GINI (good info about gini here).
- For classification the Gini cost function is used which provides an indication of how “pure” the leaf nodes are (how mixed the training data assigned to each node is).
Gini = sum(pk * (1 – pk))
- Early stop - 1 sample per node is overfitting, 5-10 are good
- Pruning - evaluate what happens if the lead nodes are removed, if there is a big drop, we need it.
- Similar to a binary search tree, just by using the median and selecting a feature randomly for each level.
- Used to find nearest neighbours.
- Many applications of using KD tree, reduce color space, Database key search, etc
How do deal with imbalanced data in Random-forest -
- One is based on cost sensitive learning.
- Other is based on a sampling technique
- A comparison between random forest and extra trees
Fig. 1: Comparison of random forests and extra trees in presence of irrelevant predictors. In blue are presented the results from the random forest and red for the extra trees. The results are quite striking: Extra Trees perform consistently better when there are a few relevant predictors and many noisy ones - Difference between RF and ET
- Differences #2