Home > Research > Publications & Outputs > Improving the Reliability for Confidence Estima...

Links

Text available via DOI:

View graph of relations

Improving the Reliability for Confidence Estimation

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
  • Haoxuan Qu
  • Yanchao Li
  • Lin Geng Foo
  • Jason Kuen
  • Jiuxiang Gu
  • Jun Liu
Close
Publication date30/10/2022
Host publicationComputer Vision – ECCV 2022 - 17th European Conference, 2022, Proceedings
EditorsShai Avidan, Gabriel Brostow, Moustapha Cissé, Giovanni Maria Farinella, Tal Hassner
Place of PublicationCham
PublisherSpringer Science and Business Media Deutschland GmbH
Pages391-408
Number of pages18
ISBN (electronic)9783031198120
ISBN (print)9783031198113
<mark>Original language</mark>English
Event17th European Conference on Computer Vision, ECCV 2022 - Tel Aviv, Israel
Duration: 23/10/202227/10/2022

Conference

Conference17th European Conference on Computer Vision, ECCV 2022
Country/TerritoryIsrael
CityTel Aviv
Period23/10/2227/10/22

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13687 LNCS
ISSN (Print)0302-9743
ISSN (electronic)1611-3349

Conference

Conference17th European Conference on Computer Vision, ECCV 2022
Country/TerritoryIsrael
CityTel Aviv
Period23/10/2227/10/22

Abstract

Confidence estimation, a task that aims to evaluate the trustworthiness of the model’s prediction output during deployment, has received lots of research attention recently, due to its importance for the safe deployment of deep models. Previous works have outlined two important qualities that a reliable confidence estimation model should possess, i.e., the ability to perform well under label imbalance and the ability to handle various out-of-distribution data inputs. In this work, we propose a meta-learning framework that can simultaneously improve upon both qualities in a confidence estimation model. Specifically, we first construct virtual training and testing sets with some intentionally designed distribution differences between them. Our framework then uses the constructed sets to train the confidence estimation model through a virtual training and testing scheme leading it to learn knowledge that generalizes to diverse distributions. We show the effectiveness of our framework on both monocular depth estimation and image classification.

Bibliographic note

Publisher Copyright: © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.