Membership Inference Attacks Against Semantic Segmentation Models

TMLR Paper878 Authors

20 Feb 2023 (modified: 05 Jun 2023)Rejected by TMLREveryoneRevisionsBibTeX
Abstract: Membership inference attacks aim to infer whether a data record has been used to train a target model by observing its predictions. In sensitive domains such as healthcare, this can constitute a severe privacy violation. In this work we attempt to address the existing knowledge gap by conducting an exhaustive study of membership inference attacks and defences in the domain of semantic image segmentation. Our findings indicate that for certain threat models, these learning settings can be considerably more vulnerable than the previously considered classification settings. We additionally investigate a threat model where a dishonest adversary can perform model poisoning to aid their inference and evaluate the effects that these adaptations have on the success of membership inference attacks. We quantitatively evaluate the attacks on a number of popular model architectures across a variety of semantic segmentation tasks, demonstrating that membership inference attacks in this domain can achieve a high success rate and defending against them may result in unfavourable privacy-utility trade-offs or increased computational costs.
Submission Length: Long submission (more than 12 pages of main content)
Assigned Action Editor: ~W_Ronny_Huang1
Submission Number: 878
Loading