Discussion of article "Advanced resampling and selection of CatBoost models by brute-force method" - page 8
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Apparently we have different ideas about random bousting. Decisive tree, it's about selected features from a random set. The point is that the sets are random, but the selection / clustering into bad good ones was originally there. It's like throwing a needle, measuring angles and calculating the number of pi)
from the wiki
I have heard about Random boosting for the first time too.
I was talking about random forest.
Yes, there are many trees, but each one is trying to train itself best on different traits. This is not the same as lumping multiple forests (including bad ones) together
However, combining case forests based on the same attributes is equivalent to 1 forest with the number of trees = the number of trees in all forests to be combined. The only difference will be different initialisation of the HCS.
Trees in a case forest are averaged.
However, merging of case forests based on the same features is equal to 1 forest with the number of trees = the number of trees in all merged forests. The only difference will be different initialisation of the HCS.
The difference is that each tree without pruning is able to perfectly remember the dataset, which causes it to retrain. An ensemble of trees is against overtraining, because some averaging occurs. But each tree is good on its own.
If you herd classifiers, it's a different story there. Averaging with a bad classifier degrades the overall result
The difference is that every tree without pruning is able to remember the dataset perfectly, which causes it to retrain. An ensemble of trees is against overlearning, since some averaging occurs. But each tree is good on its own.
If you herd classifiers, it's a different story there. Averaging with a bad classifier degrades the overall result
Besides pruning, there is a limit on depth and on the number of examples in the leaf.
A single tree is also a classifier.
I hope you will find time to compare the average and the best results on the exam sample. Not to argue theoretically, but to confirm one of the variants by practice.
I don't understand you.
This is the first time I've heard of Random boosting, too.
I was talking about random forest.
I apologise, typo. Forest of course, forest. By the way, it was first implemented in fortran 77 (with oop) in 1986, when people here were still learning fortran 4 (without oop).
But it doesn't change the point. Sampling the best features, trees in an ensemble improves the result. But at the same time clustering into good bad sets is performed from a random set of features, not the full set, which reduces the required resources, and at the same time, as practice has shown, does not significantly worsen the result.
In addition to pruning, there is a depth limit and a limit on the number of examples per sheet.
One tree is also a classifier.
I hope you will find time to compare the average and the best results on the exam sample. Not to argue theoretically, but to confirm one of the variants by practice.
Trained 20 models
Best:
All 20:
50 models
100 models
best
all
Once again, on 50 models:
Best
averages
Once again.