hello,
with code similar to the correction i got different result
from sklearn.ensemble import RandomForestClassifier
rf = RandomForestClassifier(n_estimators=300)
result_rf = cross_validate(rf, data, target,
cv=10, scoring= ['accuracy', 'balanced_accuracy'],
n_jobs=2)
from sklearn.ensemble import GradientBoostingClassifier
from sklearn.model_selection import KFold
gb = GradientBoostingClassifier(n_estimators=300)
for i in range(10):
cv = KFold(n_splits=10, shuffle=True, random_state=i )
result_rf = cross_validate(rf, data, target,
cv=cv, scoring= ['accuracy', 'balanced_accuracy'],
n_jobs=2)
result_gb = cross_validate(gb, data, target,
cv=cv, scoring= ['accuracy', 'balanced_accuracy'],
n_jobs=2)
print(result_rf['test_balanced_accuracy'].mean())
print(result_gb['test_balanced_accuracy'].mean())
print(result_gb['test_balanced_accuracy'].mean() -result_rf['test_balanced_accuracy'].mean())