dbdc3

Evaluation metrics

We use two types of evaluation metrics: classification-related and distribution-related. Here, O means not a breakdown, T possible breakdown, and X breakdown.

Classification-related metrics evaluate the accuracy related to the classification of the breakdown labels. Here, the accuracy is calculated by comparing the output of the detector and the gold label determined by majority voting. We use a threshold t to obtain the gold label.

The number of correctly classified labels divided by the total number of labels to be classified.

The precision, recall, and F-measure for the classification of the X labels.

The precision, recall, and F-measure for the classification of T + X labels; that is, PB and B labels are treated as a single label.

Distribution-related metrics evaluate the similarity of the distribution of the breakdown labels, which is calculated by comparing the predicted distribution of the labels with that of the gold labels.

Distance between the predicted distribution of the three labels and that of the gold labels calculated by Jensen-Shannon Divergence.

JS divergence when T and X are regarded as a single label.

JS divergence when O and T are regarded as a single label.

Distance between the predicted distribution of the three labels and that of the gold labels calculated by mean squared error.

Mean squared error when T and X are regarded as a single label.

Mean squared error when O and T are regarded as a single label.

How to evaluate your dialogue breakdown detector

By using the evaluation script, you can evaluate your dialogue breakdown detector by using the above evaluation metrics. Refer to this page to get information about how to use the evaluation script.