ClassifierMeasurements
ClassifierMeasurements[classifier,testset,prop]
gives measurements associated with property prop when classifier is evaluated on testset.
ClassifierMeasurements[classifier,testset]
yields a measurement report that can be applied to any property.
ClassifierMeasurements[data,…]
uses classifications data instead of a classifier.
ClassifierMeasurements[…,{prop_{1},prop_{2},…}]
gives properties prop_{1}, prop_{2}, etc.
Details and Options
 Measurements are used to determine the performance of a classifier on data that was not used for training purposes (the test set).
 Possible measurements include classification metrics (accuracy, likelihood, etc.), visualizations (confusion matrix, ROC curve, etc.) or specific examples (such as the worst classified examples).
 Classifiers can be a ClassifierFunction or a neural net (NetGraph, NetChain, etc.) that has a "Class" decoder.
 In ClassifierMeasurements[data,…], the classifications data can have the following forms:

{class_{1},class_{2},…} classifications from a classifier (human, algorithm, etc.) {dist_{1},dist_{2},…} class distributions obtained by a classifier {<class_{1}p_{1},… >,<class_{1}q_{1},… >,…} classification probabilities obtained by a classifier  ClassifierMeasurements[…,opts] specifies that the classifier should use the options opts when applied to the test set. Possible options are as given in ClassifierFunction.
 ClassifierMeasurements[classifier,testset] returns a ClassifierMeasurementsObject[…] that displays as a report panel, such as:
 ClassifierMeasurementsObject[…][prop] can be used to obtain the property prop. When repeated property lookups are required, this is typically more efficient than using ClassifierMeasurements every time.
 ClassifierMeasurementsObject[…][prop,opts] specifies that the classifier should use the options opts when applied to the test set. These options supersede original options given to ClassifierMeasurements.
 ClassifierMeasurements has the same options as ClassifierFunction[…], with the following additions:

Weights Automatic weights to be associated with test set examples ComputeUncertainty False whether measures should be given with their statistical uncertainty  When ComputeUncertaintyTrue, numerical measures will be returned as Around[result,err] where err represents the standard error (corresponding to a 68% confidence interval) associated with measure result.
 Possible settings for Weights include:

Automatic associates weight 1 with all test examples {w_{1},w_{2},…} associates weight w_{i} with the i test examples  Changing the weight of a test example from 1 to 2 is equivalent to duplicating the example.
 Weights affect measures as well as their uncertainties.
 Properties returning a single numeric value related to classification abilities on the test set include:

"Accuracy" fraction of correctly classified examples "Accuracy"n topn accuracy "AccuracyBaseline" accuracy if predicting the commonest class "CohenKappa" Cohen's kappa coefficient "Error" fraction of incorrectly classified examples "GeometricMeanProbability" geometric mean of the actualclass probabilities "LogLikelihood" loglikelihood of the model given the test set "MeanCrossEntropy" mean cross entropy over test examples "MeanDecisionUtility" mean utility over test example "Perplexity" exponential of the mean cross entropy "ScottPi" Scott's pi coefficient "RejectionRate" fraction of examples classified as Indeterminate  Examples classified as Indeterminate are discarded when measuring properties related to classification abilities on the test set, such as "Accuracy", "Error", or "MeanCrossEntropy".
 Confusion matrix–related properties include:

"ConfusionMatrix" counts c_{ij} of class i examples classified as class j "ConfusionMatrixPlot" plot of the confusion matrix "ConfusionMatrixPlot"{c_{1},c_{2},…} confusion matrix plot restricted to classes c_{1}, c_{2}, etc. "ConfusionMatrixPlot"n confusion matrix plot for the worst nclass subset "ConfusionFunction" function giving confusion matrix values "TopConfusions" pairs of classes that are most confused "TopConfusions"n n most confused class pairs  Timingrelated properties include:

"EvaluationTime" time needed to classify one example of the test set "BatchEvaluationTime" marginal time to classify one example in a batch  Properties returning one value for each testset example include:

"DecisionUtilities" value of the utility function for each example "Probabilities" actualclass classification probabilities for each example "SHAPValues" Shapley additive feature explanations for each example  "SHAPValues" assesses the contribution of features by comparing predictions with different sets of features removed and then synthesized. The option MissingValueSynthesis can be used to specify how the missing features are synthesized. SHAP explanations are given as odds ratio multipliers with respect to the class training prior. "SHAPValues"n can be used to control the number of samples used for the numeric estimations of SHAP explanations.
 Properties related to probability calibration include:

"CalibrationCurve" probability calibration curve in logit scale "LinearCalibrationCurve" probability calibration curve in linear scale "CalibrationData" probability calibration curve data  Properties returning graphics include:

"AccuracyRejectionPlot" plot of the accuracy as function of the rejection rate "ICEPlots" Individual Conditional Expectation (ICE) plots "ProbabilityHistogram" histogram of actualclass probabilities "Report" panel reporting main measurements "ROCCurve" Receiver Operating Characteristics curve for each class "SHAPPlots" Shapley additive feature explanations plot for each class  Properties returning examples from the test set include:

"Examples" all test examples "Examples"{i,j} all class i examples classified as class j "BestClassifiedExamples" examples having the highest actualclass probability "WorstClassifiedExamples" examples having the lowest actualclass probability "CorrectlyClassifiedExamples" examples correctly classified "MisclassifiedExamples" examples misclassified "TruePositiveExamples" true positive test examples for each class "FalsePositiveExamples" false positive test examples for each class "TrueNegativeExamples" true negative test examples for each class "FalseNegativeExamples" false negative test examples for each class "IndeterminateExamples" test examples classified as Indeterminate "LeastCertainExamples" examples having the highest distribution entropy "MostCertainExamples" examples having the lowest distribution entropy  Examples are given in the form input_{i}class_{i}, where class_{i} is the actual class from the test set.
 Properties such as "WorstClassifiedExamples" or "MostCertainExamples" output up to 10 examples. ClassifierMeasurementsObject[…][propn] can be used to output n examples.
 Properties returning one measure for each class include:

"AreaUnderROCCurve" area under the ROC curve for each class "ClassMeanCrossEntropy" mean cross entropy for each class "ClassRejectionRate" rejection rate for each class "F1Score" F_{1} score for each class "FalseDiscoveryRate" false discovery rate for each class "FalseNegativeRate" false negative rate for each class "FalsePositiveRate" false positive rate for each class "MatthewsCorrelationCoefficient" Matthews correlation coefficient for each class "NegativePredictiveValue" negative predictive value for each class "Precision" precision of classification for each class "Recall" recall rate of classification for each class "Specificity" specificity for each class "TruePositiveNumber" number of true positive examples "FalsePositiveNumber" number of false positive examples "TrueNegativeNumber" number of true negative examples "FalseNegativeNumber" number of false negative examples  ClassifierMeasurementsObject[…][propclass] can be used to only return the measure associated with the specified class.
 ClassifierMeasurementsObject[…][prop<class_{1}w_{1},class_{2}w_{2},… >] can be used to return a weighted average of each class measure.
 ClassifierMeasurementsObject[…][propf] can be used to apply function f to the returned class measures (e.g. ClassifierMeasurementsObject[…][propMean]).
 Properties such as "Precision" or "Recall" give one measure for each possible "positive class". The "negative class" is the union of all the classes that are not the positive class. For such properties, one can average the measures for all possible positive classes using ClassifierMeasurementsObject[…][propaverage], where average can be:

"MacroAverage" takes the mean of the measures "WeightedMacroAverage" weights each measure by its related class frequency "MicroAverage" joins true positive/true negative etc. examples of all classes to give a unique measure  Other properties include:

"ClassifierFunction" ClassifierFunction[…] being measured "Properties" list of measurement properties available
Examples
open allclose allBasic Examples (4)
Train a classifier on a training set:
Measure the accuracy of the classifier on the test set:
Visualize the confusion matrix:
Measure several properties at once:
Train a classifier on a training set:
Generate a measurement object of the classifier on a test set:
Obtain the list of measurement properties available:
Obtain the accuracy from the measurement object:
Obtain the accuracy along with its statistical uncertainty due to finite testset size:
Measure the accuracy directly from classified examples:
Measure the mean cross entropy from classification probabilities:
Measure the mean cross entropy from class distributions:
Define a neural net classifier:
Train the neural net on a training set:
Scope (7)
Decision Metrics (3)
Measure the accuracy of classified examples against their true labels:
Obtain the statistical uncertainty on this measure:
Compare the accuracy to a baseline (always predicting the most likely testset class, "B" in this case):
Measure the error on the same data:
Check that error and accuracy sum to 1:
Measure the accuracy using a higher weight for the last two examples:
Create a training set and a test set on MNIST data:
Train a model on the training set:
Compute the accuracy of the model on the test set:
Compute the "top3" accuracy of the model on the test set (classification is considered correct if the true class is among the three predicted classes with the highest probabilities):
Measure Cohen's kappa coefficient on classified examples against their true labels:
Confusion Matrix & Example Extraction (1)
Create a training set and a test set on MNIST data:
Train a model on the training set:
Create a classifier measurements object for this classifier on the test set:
Find the 10 test examples that have the worst classification (lowest probabiity for the correct class):
Compute their correctclass probability:
Find the 10 test examples that have the best classification:
Compute their correctclass probability:
Visualize the confusion matrix:
Extract the number of test examples of "3" classified as a "5":
Find the 5 most frequent class confusions:
Restrict the confusion matrix to the set of 3 classes that are the most confused with each other:
Restrict the confusion matrix to the set of 3 classes that are the least confused with each other:
Probabilistic Metrics (1)
Load the Fisher's Irises dataset:
Create a training set and a test set:
Train a classifier to recognize the iris species from their attributes:
Measure the loglikelihood of the test set (total logprobabilities of correct classes):
Measure the mean crossentropy:
The mean crossentropy is the average negative loglikelihood:
Measure the geometric mean of the probabilities for the correct class:
Probability Calibration (1)
Create a training set and a test set on MNIST data:
Train a random forest classifier for which the probability calibration is deactivated:
Visualize the reliability diagram (a.k.a calibration curve) of the classifier on the test set in a logit scale:
Visualize the same reliability diagram using a linear scaling:
Train a classifier where the calibration is activated:
BinaryClassification Metrics (1)
Load the Fisher's Irises dataset:
Create a training set and a test set:
Train a classifier to recognize the iris species from their attributes:
Compute a ROC curve for each possible "positive" class:
Compute the area under these curves:
Measure the F_{1} score for each possible "positive" class:
Measure the F_{1} score if versicolor is the "positive" class:
Take the mean of all F_{1} scores:
Weight each score according to the class prior:
Compute F_{1} scores by joining true positive/true negative, etc. examples of all classes to give a unique measure:
Options (6)
ClassPriors (1)
Load the training set and test set of the "Satellite" dataset:
Train a classifier on the training set:
Train a classifier on the training set:
Visualize the confusion matrix obtained when the classifier has a different value of ClassPriors:
Perform the same operation by first generating a ClassifierMeasurementsObject:
IndeterminateThreshold (1)
Load the training set and test set of the "Titanic" dataset:
Train a classifier on the training set:
Visualize the confusion matrix of the classifier on the test set:
Visualize the confusion matrix obtained when the classifier has a different value of IndeterminateThreshold:
Measure the accuracy of the classifier on the test set for different values of IndeterminateThreshold:
TargetDevice (1)
Train a classifier using a neural network:
Measure the accuracy of the classifier on a test set for different setting of TargetDevice:
UtilityFunction (1)
Load the training set and test set of the "Mushroom" dataset:
Train a classifier on a part of the training set:
Visualize the confusion matrix of the classifier on the test set:
Visualize the confusion matrix obtained when the classifier has a different value of UtilityFunction:
Perform the same operation by first generating a ClassifierMeasurementsObject:
ComputeUncertainty (1)
Train a classifier that classifies movie review snippets as "positive" or "negative":
Generate a ClassifierMeasurements[…] object using a test set:
Obtain a measure of the accuracy along with its uncertainty:
Obtain a measure of other properties along with their uncertainties:
Applications (2)
Train a classifier on the Fisher Iris dataset to predict the species of Iris (setosa, versicolor, virginica) from four measured features:
Measure the accuracy of the classifier on a test set:
Generate a confusion matrix to visualize the actual and predicted classifications of the test set using the classifier:
Extract examples of class "versicolor" being misclassified as "virginica":
Return the confusion matrix as a set of associations:
Train a classifier on a sample of the MNIST dataset:
Generate a ClassifierMeasurementsObject of the classifier on the MNIST test set:
Extract examples of 9 confused with 0:
Extract the 20 worst classified examples:
Compute the Fscore of each class to find for which class the classifier should be improved:
Find the minimal value for the rejection threshold in order for the accuracy to be above 95%:
Visualize the confusion matrix and compute the Fscores with this rejection threshold:
Text
Wolfram Research (2014), ClassifierMeasurements, Wolfram Language function, https://reference.wolfram.com/language/ref/ClassifierMeasurements.html (updated 2021).
CMS
Wolfram Language. 2014. "ClassifierMeasurements." Wolfram Language & System Documentation Center. Wolfram Research. Last Modified 2021. https://reference.wolfram.com/language/ref/ClassifierMeasurements.html.
APA
Wolfram Language. (2014). ClassifierMeasurements. Wolfram Language & System Documentation Center. Retrieved from https://reference.wolfram.com/language/ref/ClassifierMeasurements.html