Hello.
Why should we use KFold CV when comparing the performances of the random forest model vs. the gradient boosting one? I believe Kfold was also used in another occasion before, and i did non understand why. Is it not the basic strategy implemented in cross_validate
?
The last question concerning the BalancedBaggingClassifier
is not clear, you do not state that the base_estimator should be HistGradient. The default estimator is a decision tree, so I got obviously a worse result.