[Reproducibility Report] Explainable Deep One-Class ClassificationDownload PDF

Anonymous

05 Feb 2022 (modified: 05 May 2023)ML Reproducibility Challenge 2021 Fall Blind SubmissionReaders: Everyone
Keywords: Reproducibility, Deep Learning, Explainable AI (XAI), Anomaly Detection, Anomaly Segmentation, Pixel-wise Anomaly Detection
TL;DR: We reproduced the paper Explainable Deep One-Class Classification, reported the obtained results, and extended the results analyses
Abstract: Scope of Reproducibility Liznerski et al. [23] proposed Fully Convolutional Data Description (FCDD), an explainable version of the Hypersphere Classifier (HSC) to directly address image anomaly detection (AD) and pixel-wise AD without any post-hoc explainer methods. The authors claim that FCDD achieves results comparable with the state-of-the-art in sample-wise AD on Fashion-MNIST and CIFAR-10 and exceeds the state-of-the-art on the pixel-wise task on MVTec-AD. They also give evidence to show a clear improvement by using few (1 up to 8) real anomalous images in MVTec-AD for supervision at the pixel level. Finally, a qualitative study with horse images on PASCAL-VOC shows that FCDD can intrinsically reveal spurious model decisions by providing built-in anomaly score heatmaps. --- Methodology We have reproduced the quantitative results in the main text of [23] except for the performance on ImageNet: sample- wise AD on Fashion-MNIST and CIFAR-10, and pixel-wise AD on MVTec-AD. We used the author’s code with GPUs NVIDIA TITAN X and NVIDIA TITAN Xp. A more detailed look into FCDD’s performance variability is presented, and a Critical Difference (CD) diagram is proposed as a more appropriate tool to compare methods over the datasets in MVTec-AD. Finally, we study the generalization power of the unsupervised FCDD during training. --- Results All per-class performances (in terms of Area Under the ROC Curve (ROC-AUC) [31]) announced in the paper were replicated with absolute difference of at most 2% and below 1% on average, confirming the paper’s claims. We report the experiments’ GPU and CPU memory requirements and their average training time. Our analyses beyond the paper’s scope show that claiming to “exceed the state-of-the-art” should be considered with care, and evidence is given to argue that the pixel-wise unsupervised FCDD could narrow the gap with its semi-supervised version. --- What was easy The paper was clear and explicitly gave many training and hyperparameters details, which were conveniently set as default in the author’s scripts. Their code was well organized and easy to interact with. --- What was difficult Using ImageNet proved to be challenging due to its size and need to manually set it up; we could not complete the experiments on this dataset. --- Communication with original authors We reached the main author by e-mail to ask for help with ImageNet and discuss a few practical details. He promptly replied with useful information.
Paper Url: https://openreview.net/forum?id=A5VV3UyIQz
Paper Venue: ICLR 2021
Supplementary Material: zip
4 Replies

Loading