It could be considered a parameter of the model but not a hyperparameter: you will not change this threshold during the fit call. But you can post-tune the decision once you got the model.
Right. To be even clearer, I guess one should treat Precision/Recall and ROC curves as “metrics” for obtaining the “best” model overall. But then, given those curves at a certain point in the development, is it correct to choose the probability threshold that maximizes the precision/recall trade-off? This would mean that one could be forced to choose different values for the probability threshold along the model evolution, right?
This is not currently available in scikit-learn.
So how does one set and use the probability threshold in predict
when using scikit-learn? Simple python code like: if classifier.predict_proba(data) > choosen_threshold ...
?