CALIBRATING CONFIDENCE OF CLASSIFICATION MODELS
Abstract:
Disclosed is a technical solution to calibrate confidence scores of classification networks. A classification network has been trained to receive an input and output a label of the input that indicates a class of the input. The classification network also outputs a confidence score of the label, which indicates a likelihood of the input falling into the class, i.e., a confidence level of the classification network that the label is correct. To calibrate the confidence of the classification network, a logit transformation function may be added into the classification network. The logic transformation function may be an entropy-based function and have learnable parameters, which may be trained by inputting calibration samples into the classification network and optimizing a negative log likelihood based on the labels generated by the classification network and ground-truth labels of the calibration samples. The trained logic transformation function can be used to compute reliable confidence scores.
Information query
Patent Agency Ranking
0/0