Presentation + Paper
7 June 2024 Extracting explanations, justification, and uncertainty from black-box deep neural networks
Paul Ardis, Arjuna Flenner
Author Affiliations +
Abstract
Deep Neural Networks (DNNs) do not inherently compute or exhibit empirically-justified task confidence. In mission critical applications, it is important to both understand associated DNN reasoning and its supporting evidence. In this paper, we propose a novel Bayesian approach to extract explanations, justifications, and uncertainty estimates from DNNs. Our approach is efficient both in terms of memory and computation, and can be applied to any black box DNN without any retraining, including applications to anomaly detection and out-of-distribution detection tasks. We validate our approach on the CIFAR-10 dataset, and show that it can significantly improve the interpretability and reliability of DNNs.
Conference Presentation
(2024) Published by SPIE. Downloading of the abstract is permitted for personal use only.
Paul Ardis and Arjuna Flenner "Extracting explanations, justification, and uncertainty from black-box deep neural networks", Proc. SPIE 13054, Assurance and Security for AI-enabled Systems, 1305405 (7 June 2024); https://doi.org/10.1117/12.3012765
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Covariance

Neural networks

Data modeling

Machine learning

Artificial intelligence

Matrices

Safety

Back to Top