The adoption of deep learning models has brought significant performance improvements across several research fields, such as computer vision and natural language processing. However, their 'black-box' nature yields the downside of poor explainability: in particular, several real-world applications require - to varying extents - reliable confidence scores associated to a model's prediction. The relation between a model's accuracy and confidence is typically referred to as calibration. In this work, we propose a novel calibration method based on gradient accumulation in conjunction with existing loss regularization techniques. Our experiments on the Named Entity Recognition task show an improvement of the performance/calibration ratio compared to the current methods.
A Novel Gradient Accumulation Method for Calibration of Named Entity Recognition Models / Jouet, G.; Duhart, C.; Staiano, J.; Rousseaux, F.; De Runz, C.. - 2022-:(2022), pp. 01-08. (Intervento presentato al convegno 2022 International Joint Conference on Neural Networks, IJCNN 2022 tenutosi a Padova nel 2022) [10.1109/IJCNN55064.2022.9892324].
A Novel Gradient Accumulation Method for Calibration of Named Entity Recognition Models
Staiano J.;
2022-01-01
Abstract
The adoption of deep learning models has brought significant performance improvements across several research fields, such as computer vision and natural language processing. However, their 'black-box' nature yields the downside of poor explainability: in particular, several real-world applications require - to varying extents - reliable confidence scores associated to a model's prediction. The relation between a model's accuracy and confidence is typically referred to as calibration. In this work, we propose a novel calibration method based on gradient accumulation in conjunction with existing loss regularization techniques. Our experiments on the Named Entity Recognition task show an improvement of the performance/calibration ratio compared to the current methods.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione