Causes and Effects of Unanticipated Numerical Deviations in Neural Network Inference Frameworks

Published: 21 Sept 2023, Last Modified: 22 Dec 2023NeurIPS 2023 posterEveryoneRevisionsBibTeX
Keywords: machine learning, security, reproducibility, forensics
TL;DR: We measure how different hardware affects the results of neural network inference while keeping the trained model and the input constant.
Abstract: Hardware-specific optimizations in machine learning (ML) frameworks can cause numerical deviations of inference results. Quite surprisingly, despite using a fixed trained model and fixed input data, inference results are not consistent across platforms, and sometimes not even deterministic on the same platform. We study the causes of these numerical deviations for convolutional neural networks (CNN) on realistic end-to-end inference pipelines and in isolated experiments. Results from 75 distinct platforms suggest that the main causes of deviations on CPUs are differences in SIMD use, and the selection of convolution algorithms at runtime on GPUs. We link the causes and propagation effects to properties of the ML model and evaluate potential mitigations. We make our research code publicly available.
Supplementary Material: pdf
Submission Number: 10954
Loading