Vision Transformer Reliability Evaluation on the Coral Edge TPU

Bruno Loureiro Coelho, Pablo Rafael Bodmann, Niccolò Cavagnero, Christopher Frost, Paolo Rech

Published: 01 Apr 2025, Last Modified: 16 Feb 2026IEEE Transactions on Nuclear ScienceEveryoneRevisionsCC BY-SA 4.0
Abstract: Vision transformers (ViTs) outperform convolutional neural networks (CNNs) in tasks such as image classification, and, despite their high computational complexity, they can still be mapped to low-power EdgeAI accelerators, such as the Coral tensor processing unit (TPU). In this article, through accelerated neutron beam experiments, we study the reliability of six ViTs on the Coral TPU and four microbenchmarks. According to our data, the internal size of attention heads (the main computational block in ViTs) has negligible impact on the failure-in-time (FIT) rate of the model compared to increasing the number of heads in the model; furthermore, our results show that employing convolutions in the patch embedding reduces the FIT rate of the model. Additionally, we decompose ViTs into four basic computational blocks that represent the main operators of the model, showing that, although the transformer layer [with multihead self-attention and multilayer perceptron (MLP)] presents the highest FIT rate, it is actually the patch embedding that is more likely to cause misclassifications. These results can be leveraged to design hardening techniques that improve the resilience of the critical blocks of a ViT, identified in our evaluation while minimizing the additional overhead.
Loading