Mena, JoséPujol Vila, OriolVitrià i Marca, Jordi2020-07-142020-07-142020-05-212169-3536https://hdl.handle.net/2445/168537Machine Learning as a Service platform is a very sensible choice for practitioners that wantto incorporate machine learning to their products while reducing times and costs. However, to benefit theiradvantages, a method for assessing their performance when applied to a target application is needed. In thiswork, we present a robust uncertainty-based method for evaluating the performance of both probabilistic andcategorical classification black-box models, in particular APIs, that enriches the predictions obtained withan uncertainty score. This uncertainty score enables the detection of inputs with very confident but erroneouspredictions while protecting against out of distribution data points when deploying the model in a productivesetting. We validate the proposal in different natural language processing and computer vision scenarios.Moreover, taking advantage of the computed uncertainty score, we show that one can significantly increasethe robustness and performance of the resulting classification system by rejecting uncertain predictions26 p.application/pdfengcc-by (c) Mena, José et al., 2020http://creativecommons.org/licenses/by/3.0/esAprenentatge automàticSistemes classificadors (Intel·ligència artificial)Intel·ligència artificialMachine learningLearning classifier systemsArtificial intelligenceUncertainty-based Rejection Wrappers for Black-box Classifiersinfo:eu-repo/semantics/article7007932020-07-14info:eu-repo/semantics/openAccess