We introduce algorithmic assurance, the problem of testing whether machine learning algorithms are conforming to their intended design goal. We address this problem by proposing an efficient framework for algorithmic testing. To provide assurance, we need to efficiently discover scenarios where an algorithm decision deviates maximally from its intended gold standard. We mathematically formulate this task as an optimisation problem of an expensive, black-box function. We use an active learning approach based on Bayesian optimisation to solve this optimisation problem. We extend this framework to algorithms with vector-valued outputs by making appropriate modification in Bayesian optimisation via the EXP3 algorithm. We theoretically analyse our methods for convergence. Using two real-world applications, we demonstrate the efficiency of our methods. The significance of our problem formulation and initial solutions is that it will serve as the foundation in assuring humans about machines making complex decisions.
History
Volume
31
Pagination
1-9
Location
Montreal, Canada
Start date
2018-12-02
End date
2018-12-08
ISSN
1049-5258
Language
eng
Publication classification
E1 Full written paper - refereed
Copyright notice
2018, Neural Information Processing Systems Foundation, Inc.
Editor/Contributor(s)
Bengio S, Wallach H, Larochelle H, Grauman K, CesaBianchi N, Garnett R
Title of proceedings
NeurIPS 2018 : Proceedings of the 32nd Conference on Neural Information Processing Systems
Event
Neural Information Processing Systems Foundation. Conference (32nd : 2018 : Montreal, Canada)
Publisher
Neural Information Processing Systems Foundation
Place of publication
San Diego, Calif.
Series
Neural Information Processing Systems Foundation Conference