6 research outputs found

    Explaining Convolutional Neural Networks through Attribution-Based Input Sampling and Block-Wise Feature Aggregation

    Full text link
    As an emerging field in Machine Learning, Explainable AI (XAI) has been offering remarkable performance in interpreting the decisions made by Convolutional Neural Networks (CNNs). To achieve visual explanations for CNNs, methods based on class activation mapping and randomized input sampling have gained great popularity. However, the attribution methods based on these techniques provide lower resolution and blurry explanation maps that limit their explanation power. To circumvent this issue, visualization based on various layers is sought. In this work, we collect visualization maps from multiple layers of the model based on an attribution-based input sampling technique and aggregate them to reach a fine-grained and complete explanation. We also propose a layer selection strategy that applies to the whole family of CNN-based models, based on which our extraction framework is applied to visualize the last layers of each convolutional block of the model. Moreover, we perform an empirical analysis of the efficacy of derived lower-level information to enhance the represented attributions. Comprehensive experiments conducted on shallow and deep models trained on natural and industrial datasets, using both ground-truth and model-truth based evaluation metrics validate our proposed algorithm by meeting or outperforming the state-of-the-art methods in terms of explanation ability and visual quality, demonstrating that our method shows stability regardless of the size of objects or instances to be explained.Comment: 9 pages, 9 figures, Accepted at the Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21

    Visual Post-hoc Explanation of Convolutional Neural Networks Via Attribution-based Perturbation

    No full text
    Visual explanation algorithms are a subset of post-hoc solutions in Explainable AI that aim to open the black-box of cumbersome models utilized for image recognition tasks. These algorithms aim to determine the features in each given input data that contribute the most to the model's prediction. Producing high-resolution and class-discriminative visual explanations that accurately represent the evidence leading the model to make its predictions is a common challenge in prior visual explanation solutions. This thesis proposes a novel CNN-specific visual explanation algorithm that circumvents these limitations by extracting and aggregating semantic and spatial information from various layers of CNN. The performance, complexity, and compatibility of our proposed algorithm are analyzed through thorough comparisons with state-of-the-art methods in large-scale experiments on natural scene object recognition and visual anomaly inspection applications. Furthermore, this thesis introduces a small-scale benchmark for validating the performance of visual explanation methods on a pattern classification task.M.A.S

    Recommendations from the ICM-VTE: General

    No full text
    corecore