spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Apache Spark (JIRA)" <>
Subject [jira] [Commented] (SPARK-6332) compute calibration curve for binary classifiers
Date Sat, 14 Mar 2015 02:20:38 GMT


Apache Spark commented on SPARK-6332:

User 'robert-dodier' has created a pull request for this issue:

> compute calibration curve for binary classifiers
> ------------------------------------------------
>                 Key: SPARK-6332
>                 URL:
>             Project: Spark
>          Issue Type: New Feature
>          Components: MLlib
>            Reporter: Robert Dodier
>            Priority: Minor
>              Labels: classification
> For binary classifiers, calibration measures how classifier scores compare to the proportion
of positive examples. If the classifier is well-calibrated, the classifier score is approximately
equal to the proportion of positive examples. This is important if the scores are used as
probabilities for making decisions via expected cost. Otherwise, the calibration curve may
still be interesting; the proportion of positive examples should at least be a monotonic function
of the score.
> I propose that a new method for calibration be added to the class BinaryClassificationMetrics,
since calibration seems to fit in with the ROC curve and other classifier assessments. 
> For more about calibration, see:
> References:
> Mahdi Pakdaman Naeini, Gregory F. Cooper, Milos Hauskrecht. "Binary Classifier Calibration:
Non-parametric approach."
> Alexandru Niculescu-Mizil, Rich Caruana. "Predicting Good Probabilities With Supervised
Learning." Appearing in Proceedings of the 22nd International Conference on Machine Learning,
Bonn, Germany, 2005.
> "Properties and benefits of calibrated classifiers." Ira Cohen, Moises Goldszmidt.

This message was sent by Atlassian JIRA

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message